StableVideo
StableVideo copied to clipboard
[ICCV 2023] StableVideo: Text-driven Consistency-aware Diffusion Video Editing
In Currently readme.md **,the origin project structure** was wrong ``` StableVideo ├── ... ├── ckpt │ ├── cldm_v15.yaml | ├── dpt_hybrid-midas-501f0c75.pt │ ├── control_sd15_canny.pth │ └── control_sd15_depth.pth ├── data │...
The provided requirements.txt for the StableVideo model lacks crucial dependencies and contains incorrect version specifications, leading to installation failures. Many essential packages required for the model's proper functioning are missing,...
The CPU only version errors out on running as the code still believes it is running on a CUDA compatible device and not on the CPU. Similarly, on opening the...
On running repo with the default configuration, we get the following exception. It doesn't interrupt the working yet it's raised. 
While running the Gradio app using the default repository configuration and data provided, the ```TypeError: Cannot interpret 'torch.float32' as a data type``` arises when loading the video. What could possibly...
Thanks for the great work. When editing a new video based on NLA models, we should first extract the frames within a video and save it into a `video_name/frames_xxx.jpg`. I...
I want to modify the ControlNet condition. But I have no idea how to train the guidance diffusion.
Loaded model config from [ckpt/cldm_v15.yaml] Loaded state_dict from [ckpt/control_sd15_canny.pth] /home/xufang/ProgramFiles/anaconda3/envs/stablevideo/lib/python3.11/site-packages/timm/models/_factory.py:114: UserWarning: Mapping deprecated model name vit_base_resnet50_384 to current vit_base_r50_s16_384.orig_in21k_ft_in1k. model = create_fn( Traceback (most recent call last): File "/home/xufang/StableVideo/app.py", line...
Hi, I'm not a pro dev and I have to say that the readme is quite ambiguous for example how to install Controlnet, just some generic link where it doesn't...