-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU memory requirement for inference #4
Comments
it requires more than 40GB for 2 seconds of 720p video in my early experiments, 3 seconds video needs ~71 GB Vram without upscaling (upscale = 1) Another question is "can we use something like flash attention to reduce vram usage"? |
At this time, A100 80G machine is required for high-resolution (~2k) and high-frame-rate (fps>=24) video generation. You can decline "up_scale" or "target_fps" to avoid OOM, but the visual results will observe obvious drop. |
You are correct, this algorithm is expensive. Actually, we have already incorporated xformers for attention computation. |
2 秒的 720p 视频用了多长时间推理? @JC1DA |
Hi, thank you so much for open sourcing this amazing work!
I'm wondering what's the memory requirement to run the inference script? I tested the script verbatim on an A100 40G machine and it went OOM. Curious if we need to use a 80G machine instead, or is there something obvious that I'm missing?
Thanks!
The text was updated successfully, but these errors were encountered: