video-diffusion-pytorch
video-diffusion-pytorch copied to clipboard
Training Dataset
Hi, thanks for this fantastic work!
Could I ask what is the training dataset for text-to-video generation? And any training code provided for this task?
Thank you so much
Looking forward to the dataset as well. If this dataset cannot be released now, could you kindly tell us how does this dataset constructed, e.g., the captions are manually labeled or extracted from pretrained networks?
Anyone has an idea what kind of public dataset we can use to test except the moving MNIST? The public dataset should also contain the respective annotations.