Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CUDA Out of Memory with llama3_text2cypher_chat.ipynb" #4

Open
mathysferrato opened this issue May 30, 2024 · 0 comments
Open

CUDA Out of Memory with llama3_text2cypher_chat.ipynb" #4

mathysferrato opened this issue May 30, 2024 · 0 comments

Comments

@mathysferrato
Copy link

mathysferrato commented May 30, 2024

Hi,

On the first run of the notebook "llama3_text2cypher_chat.ipynb" i got a CUDA Out of memory error when executing the cell with trainer.train()

I use it locally on a PC with a RTX 2070 (8GB of VRAM) and I use a quantized version with setting "load_in_4bit" to True in the first cells.
Only loading the model when doing "model, tokenizer = FastLanguageModel.from_pretained(...) already consumes 6GB of VRAM, even the 4bit quantized one.

Is it normal ? I really need to launch it on my machine and not on colab.

Thanks,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant