Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How much vram needed to finetune 3b model? Is 12gb enough? #15

Open
universewill opened this issue Sep 28, 2023 · 1 comment
Open

How much vram needed to finetune 3b model? Is 12gb enough? #15

universewill opened this issue Sep 28, 2023 · 1 comment

Comments

@universewill
Copy link

How much vram needed to finetune 3b model? Is 12gb enough?

@CStanKonrad
Copy link
Owner

Unfortunately, 12GB is not enough to finetune the 3B model in the standard way (tuning all parameters). That is because of optimizer variables and values for gradient accumulation. This Hugging Face blog post briefly describes how much each of those parts contributes to VRAM usage.
For our model, we have used a single A100 80GB GPU and usage metrics show that > 70GB of the GPU memory was allocated.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants