LlamaGen icon indicating copy to clipboard operation
LlamaGen copied to clipboard

Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation

Results 69 LlamaGen issues
Sort by recently updated
recently updated
newest added

May I ask why VQVAE's learned codebook was not used as the Embedding layer for autoregressive training of language models, but only the indices of the codebook were used, and...

Hi, in the paper, you state "The codebook usage is calculated as the percentage of used codes in the queue of size 65536 over the whole codebook size." May I...

Hello, the paper mentions using Llama for autoregressive training, but why is the language model in the code using GPT ?

When I open the `https://huggingface.co/spaces/FoundationVision/LlamaGen`, It is showed as follow: runtime error Traceback (most recent call last): File "/home/user/app/app.py", line 171, in t5_model, vq_model, gpt_model, image_size = load_model(args) File "/home/user/app/app.py",...

Hi, thanks for you job, could you show us how your model performs on MS-COCO?

Hi~ Thank you for your work. I would like to know if you have compared the ablation experiments on the LAION-COCO fine-tuning.

Hi, I am trying to finetune the decoder only of the tokenizer on a new dataset. I was wondering if you could share some finetuning recipies. Mainly: - Should the...

Hello! Thank you for the clean + user friendly codebase! I'm trying to finetune the VQ-VAE tokenizer and noticed some keys might be missing from the pretrained checkpoint listed on...

Hello, due to limited computing resources, it is difficult for me to reproduce Llamagen from the beginning of training. If I only need to perform inference fine-tuning, what should I...

Thanks for the amazing work, could you open the training cost for each model? such as training GPU times and the least GPU needed.