Frequently Asked Questions #1013
-
Please submit questions here that will likely affect a lot of other users and that haven't been sufficiently covered in the documentation. |
Beta Was this translation helpful? Give feedback.
Replies: 6 comments 10 replies
-
How to implement/add a new optimizer? Because I would really love to have AdaBelief8bit 😄 |
Beta Was this translation helpful? Give feedback.
-
Hi, I was wondering if |
Beta Was this translation helpful? Give feedback.
-
Hi Team, when i am running the above qlora code for owl-vit model (google/owlvit-base-patch32) with below 4 bits bnbconfig , the fine tuning is taking place without any error. bnb_config = BitsAndBytesConfig( once i change the config with below information bnb_config = BitsAndBytesConfig( i receive the following error trace. ######################################################################################### File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/transformers/trainer.py:1537, in Trainer.train(self, resume_from_checkpoint, trial, ignore_keys_for_eval, **kwargs) File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/transformers/trainer.py:1854, in Trainer._inner_training_loop(self, batch_size, args, resume_from_checkpoint, trial, ignore_keys_for_eval) File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/transformers/trainer.py:2744, in Trainer.training_step(self, model, inputs) File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/accelerate/accelerator.py:1907, in Accelerator.backward(self, loss, **kwargs) File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/torch/_tensor.py:492, in Tensor.backward(self, gradient, retain_graph, create_graph, inputs) File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/torch/autograd/init.py:251, in backward(tensors, grad_tensors, retain_graph, create_graph, grad_variables, inputs) File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/torch/autograd/function.py:288, in BackwardCFunction.apply(self, *args) File ~/miniconda3/envs/testenv/lib/python3.10/site-packages/bitsandbytes/autograd/functions.py:491, in MatMul8bitLt.backward(ctx, grad_output) RuntimeError: The size of tensor a (32) must match the size of tensor b (4) at non-singleton dimension 0 ####################################################################################### thanks. |
Beta Was this translation helpful? Give feedback.
-
Few suggests for an FAQ:
|
Beta Was this translation helpful? Give feedback.
-
I am trying to get bitsandbytes compiled for Windows 11 with CUDA. I am using the specified versions of the Microsoft compiler (17.9.6) and CUDA (12.3). The last step in the documentation (https://huggingface.co/docs/bitsandbytes/main/en/installation?source=Windows) says to run: However, there is no build.py in the repo. I have tried running the setup.py, but having various difficulties, but suspect the problem is I am doing something else wrong. BTW, the build had several warnings, not clear if any of them are fatal. Sorry if this too elementary for this forum... |
Beta Was this translation helpful? Give feedback.
-
Hello! I want to know if bnb support deepspeed zero3_init? #1249 |
Beta Was this translation helpful? Give feedback.
Tim gave the following feedback on this question: