feed_forward_vqgan_clip icon indicating copy to clipboard operation
feed_forward_vqgan_clip copied to clipboard

Feed forward VQGAN-CLIP model, where the goal is to eliminate the need for optimizing the latent space of VQGAN for each input prompt

Results 9 feed_forward_vqgan_clip issues
Sort by recently updated
recently updated
newest added

For lack of a better word; I've noticed during training that the VitGAN tends to get stuck on one, two, or three (i don't see four happen very often/at all)...

Hi, First of all great work! I really loved it. To replicate, I tried training on the Conceptual 12M Dataset with the depth and dims same as the pretrained models...

Thanks for your excellent repo. When training `cc12m_32x1024` with type `VitGAN` or `MLP Mixer`, what kinds of GPU environment do you use? Tesla V100 with 32G mem or others? Thanks

It's quite easy to finetune one of the Open AI CLIP checkpoints with this codebase: https://github.com/Zasder3/train-CLIP-FT Uses pytorch-lightning. May be worth pursuing

Hi @mehdidc 👋🏼 I'm a new team member at @replicate. I was [trying out your model on replicate.ai](https://replicate.ai/mehdidc/feed_forward_vqgan_clip) and noticed that the names of the models are a bit cryptic,...

Hi! I really love this idea and think that this concept solves the main bottleneck of current VQGAN+CLIP approach which is the optimisation for each prompt. I love how instantaneous...

Katherine has released a better notebook for the CLIP-guided-diffusion. Outputs on a P100 is quite slow; but results can be very good. I've put the new notebook in my current...

are you familiar with this https://twitter.com/e08477/status/1418440857578098691?s=21 ? I want to do cityscape shots. Are you familiar with any relevant datasets? Can this repo help output higher quality images? Or does...

I've been generating images using this model, which is delightfully fast, but I've noticed that it produces images that are all alike. I tried generating the "null" image by doing:...