Distilled 9.6How many VRAMs are needed and how long can a video be generated
Distilled 9.6How many VRAMs are needed and how long can a video be generated
Not gonna work with 6gb of vram for sure
works with 8. just do lower res and start with something like 320x240 with 15fps target with 33 frames , use tiled vae with tile size lower than 512.
Not gonna work with 6gb of vram for sure
In my experience 6GB is totally fine though. Been able to do more than 100 frames with 480p resolution. Perhaps ComfyUI's memory management is better than this repo's, but in a sense ComfyUI is also an official implementation (the plugin is fully maintained by LightTricks themselves).
Not gonna work with 6gb of vram for sure
In my experience 6GB is totally fine though. Been able to do more than 100 frames with 480p resolution. Perhaps ComfyUI's memory management is better than this repo's, but in a sense ComfyUI is also an official implementation (the plugin is fully maintained by LightTricks themselves).
You're right, it actually runs quite well with Comfyui on just 6gb of VRAM, which honestly surprised me. Was expecting the opposite.
Even 128x 128 x 30 is causing OOM on a 24GB A5000. I think without ComfyUI it's impossible to run locally. Maybe a card with >32GB can work
I guess that most of the VRAM burden is on text encoding with those heavy t5 models (on my Comfy setup using the vanilla checkpoint without any memory optimization tricks it pretty much never OOM on my 6gb card). Not pretty sure if the official repo automatically unload them when not in use (which is essential for optimizing memory use), but on the other hand ComfyUI supports quantized clip models, which really makes a huge difference on VRAM requirement.
GTX 1650 Mobile 4GB Vram 32 GB Ram 768x512 resolution, 49 frames, 24 fps, 8 steps, with google_t5-v1_1-xxl_encoderonly-fp16 , Generation taking 1 minute exactly. Times double when double dimension or double frames. It was possible to make 1024x768, 129 frames video in roughly 10 minutes. Extremely impressive, time taken to generate initial image is longer. Black magic.
Even 128x 128 x 30 is causing OOM on a 24GB A5000. I think without ComfyUI it's impossible to run locally. Maybe a card with >32GB can work
I havent been able to get this to run just using python inference, 512x512, 128 frames, 24 fps. I run out of memory on a 32G RTX 5090 even though 31G was free. I havent used ComfyUI before. I was looking to build a workflow using just python.
Using Distilled 9.6 on 8 GB VRAM, with high resolution and 249 frames. takes total 4 minutes. Should work on 6 GB.