-
Notifications
You must be signed in to change notification settings - Fork 55
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
VRAM requirements? #1
Comments
I2VGen-XL-based I try to run it on 24 and get out of memory. |
Regarding this, I would also like to guide with specific parameters. Currently, inference on the V100 seems to be running into out-of-memory issues. Is there support for FP16 or INT8? Best regards. |
Thank you for your interest in our work! Regarding the VRAM requirements, with the default settings and the toy example we provided, the GPU peak memory usage is approximately 39GB. Currently, there are two ways to reduce the VRAM requirements: (1) decrease the frame_length; and (2) decrease the chunk_size. |
Hi @CSRuiXie,
Still has OOM error, how to fix that? |
Hi, I believe the main issue is that your input video resolution is too large for 4x upscaling. For example, with the default settings, upscaling a 640x480 video by 4x can require more than 80GB of VRAM. |
Numbers are huge any way to quantize, tile or slice and reduce VRAM? |
Yes, we are aware of the VRAM issue, and we definitely plan to introduce some techniques to optimize it, such as tiling. In the meantime, you can follow this instruction to reduce VRAM usage. |
@CSRuiXie can you modify app here and add these 2 options? https://huggingface.co/spaces/SherryX/STAR/blob/main/app.py ty so much |
Thank you for sharing your work with us. |
yes this needs to be made so that longer videos can be processed |
I also did a quick test using sample video (023_klingai_reedit.mp4) on 8 GB VRAM + 8 GB shared. Earlier was getting OOM on 1st step but after making few changes it start to work but slow as h...
I guess frame-wise processing should make it work on atleast 16/24GB VRAM. Any dirty/quick code to try frame-wise processing. |
We have now added these two options to the STAR demo. You can visit the Hugging Face demo to check the updates. |
Thank you for your interest in our work. You can set the frame_length to 1 for frame-wise processing. However, the restored results may be worse than the default setting, mainly due to two reasons: (1) the frame length during inference differs significantly from that during training, and (2) the model cannot extract temporal information from other frames. |
Please put minimum VRAM in the install instructions. |
Thanks for your advice! We will add more details about the VRAM requirements in the installation instructions. |
Thank you for the detailed explanation on VRAM management. I have a follow-up question: Would your script work with two RTX 3090 GPUs using DistributedDataParallel (DDP)? Model Parallelism ? |
Hello, first of all thank you for this awesome model. I wanted to ask what the VRAM requirements are, since I tried to run the I2VGen-XL-based model and it OOM with an RTX 4090 24gb.
The text was updated successfully, but these errors were encountered: