You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm able to get the generate.py script working. However, with the finetune.py script I'm facing the following error:
It seems to be because the load_in_8bit parameter is set to True and it's looking for a quantisation_config,json but if I set it to False then even a GPU with vRAM of 44Gb is not enough to train the model.
How do I create the quantisation_config,json? I'm using huggyllama/llama-7b as the base model since the given link for the base model is down. I face the same error when I use baffo32/decapoda-research-llama-7B-hf as the base model.
Any help would be appreciated, thank you!
The text was updated successfully, but these errors were encountered:
Nimisha-Pabbichetty
changed the title
Out of memory error with 44Gb VRAM in my GPU
Load_in_8bit causing issues: Out of memory error with 44Gb VRAM in my GPU or device_map error
Nov 11, 2023
I'm able to get the generate.py script working. However, with the finetune.py script I'm facing the following error:
It seems to be because the load_in_8bit parameter is set to True and it's looking for a quantisation_config,json but if I set it to False then even a GPU with vRAM of 44Gb is not enough to train the model.
How do I create the quantisation_config,json? I'm using huggyllama/llama-7b as the base model since the given link for the base model is down. I face the same error when I use baffo32/decapoda-research-llama-7B-hf as the base model.
Any help would be appreciated, thank you!
The text was updated successfully, but these errors were encountered: