Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Minimum VRAM requirements? #39

Open
frutiemax92 opened this issue Jan 10, 2025 · 1 comment
Open

Minimum VRAM requirements? #39

frutiemax92 opened this issue Jan 10, 2025 · 1 comment

Comments

@frutiemax92
Copy link

I've tried to run the inference script with a few tweaks to the python code (such as not loading the weights on the gpu twice), but I still cannot run it in bfloat16 without an OOM error. I have a RTX4070 (12GB VRAM): what are the minimum requirements for running inference with this model and are there any plan for optimizing the VRAM usage? Thanks.

@hrz2000
Copy link

hrz2000 commented Jan 17, 2025

Is 32GB the minimum requirement? @iFighting

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants