hht2001

Results 22 comments of hht2001

> Dear authors, thanks for your interesting work and plans. However, there is one question in my mind: why you choose to use VQVAE instead of VAE? As stated both...

准确的相机位姿已经可以靠ControlNet去实现了

> > Hello, does this error occur when you want to train a model? Because this script allows you to train from scratch without a pre-trained model. Or does it...

> For what I understand it's not a controlnet but a connector used for the controlnet used in [SUPIR](https://github.com/Fanghua-Yu/SUPIR): > > > It designs an effective 620 million parameter adaptor...

> For what I understand it's not a controlnet but a connector used for the controlnet used in [SUPIR](https://github.com/Fanghua-Yu/SUPIR): > > > It designs an effective 620 million parameter adaptor...

You can find "torch.abs(pixel_prop[:,:1,:,:] - curr_lr[:,:1,:,:])" in RNN_Swin_4.py, Line 216&239. The "1" means only use the first dimension because all the dimension are same(if use the gray input). You can...

> u can process in this way: if u have a rgb image "217, 65, 43" for "r, g, b", then u can get three "gray image"-----"217, 217, 217" ,...

hi, I commit a test_video.py . u can use it to process ur own data. waiting for author's check

> @KhawlahB I am trying similar things, how about your results? Relly looking forward to your response:) while compute attention in the unet, the prompt embedding will be the key...