Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 394.00 MiB #23

Open
shifu-learner opened this issue Jan 16, 2024 · 3 comments

Comments

@shifu-learner
Copy link

Hello,
I am trying to finetune GPT-j-6b.
I followed the instructions provided in the documentation. But, I get this error.

I tried by changing batch size =1, gradient_accumulation_steps=4.

Any idea how can i solve this.

@mallorbc
Copy link
Owner

Need more information. How are you running it? What is your hardware?

@shifu-learner
Copy link
Author

Sure
I am trying to execute in a TESLA VM with Nvidia driver.
I have followed the documentation to finetune GPT_J_6b.

@shavingtonpitsos
Copy link

is sounds like your GPU is not large enough, how much GPU ram is available to you when you run your script? It sounds like you likely have too little GPU ram, probably less than 16 GB

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants