Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Load_in_8bit causing issues: Out of memory error with 44Gb VRAM in my GPU or device_map error #604

Open
Nimisha-Pabbichetty opened this issue Nov 11, 2023 · 1 comment

Comments

@Nimisha-Pabbichetty
Copy link

I'm able to get the generate.py script working. However, with the finetune.py script I'm facing the following error:
image

It seems to be because the load_in_8bit parameter is set to True and it's looking for a quantisation_config,json but if I set it to False then even a GPU with vRAM of 44Gb is not enough to train the model.
How do I create the quantisation_config,json? I'm using huggyllama/llama-7b as the base model since the given link for the base model is down. I face the same error when I use baffo32/decapoda-research-llama-7B-hf as the base model.

Any help would be appreciated, thank you!

@Nimisha-Pabbichetty Nimisha-Pabbichetty changed the title Out of memory error with 44Gb VRAM in my GPU Load_in_8bit causing issues: Out of memory error with 44Gb VRAM in my GPU or device_map error Nov 11, 2023
@Minimindy
Copy link

I think is run out off memory, maybe you should try colab or clean up memory for space to load model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants