-
Notifications
You must be signed in to change notification settings - Fork 203
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
very very slowly RTX4090 #41
Comments
The current version of |
@bssrdf - I think the more impactful aspect of that change is converting the UNET model to bf16. That shaves a large amount off of the memory footprint and has negligible impact on the output quality. I didn't care for loading and unloading the model - it substantially reduced the memory footprint, again, but added unwanted reloads that had more of an impact on overall speed for me than was gained by freeing up GPU RAM. (YMMV naturally) |
yes, |
Not sure if this repo will receive this much of an optimization, but if what you need to fix is only about T5, there are quantized version of it that significantly reduce the requirements down to (I guess) 6GB or less. I have been able to generate at 768x512x97 on 6GB VRAM (+16GB RAM, a pretty mediocre spec). I have written instruction on how to do so on ComfyUI at #4 (comment). If you really want to try it out now on your machine, I believe this is currently the only way to go. |
--height 704 --width 480 --num_frames 121 RTX4090 need 2.5hours???
The text was updated successfully, but these errors were encountered: