Text-To-Video-Finetuning
Text-To-Video-Finetuning copied to clipboard
Finetune ModelScope's Text To Video model using Diffusers 🧨
when running inference with an init_video, runtime error happen. timesteps = timesteps[0] (inference.py line194) → for timestep in timesteps(diffusers/schedulers/scheduling_dpmsolver_multistep.py line900) This error, "iteration over a 0-d tensor," occurs when attempting...
Evals
added CLIP and FID evals and interval and random frame sampling.
Hi, I have been trying to do fine-tuning with stable LoRA, according to the manual. I only can do the basics, so I haven't modified the stable_lora_config.yaml other than the...
Are there any plans to add ControlNet? Or is it possible to use this model with ControlNet Pipeline from diffusers?
Is there a way I can set the train config to do a normal finetuning on a large dataset instead of LoRA?
please help me  
https://github.com/damo-vilab/videocomposer Modelscope and videocomposer both seem to come from Alibaba.
Hey, thanks for open-sourcing this code! I had a quick question about the `finetune_unet` function in `train.py`: why are there two forward passes and loss computations through the unet? Is...
How to enable multi-GPU training? No matter how many GPUs I use, only one process starts.
When trying to run inference using --lora_path parameter, getting : ``` LoRA rank 64 is too large. setting to: 4 list index out of range Couldn't inject LoRA's due to...