Tianhong Li

Results 65 comments of Tianhong Li

It only uses training set.

Change the dataset here to your customized dataset implementation https://github.com/LTH14/mage/blob/main/main_pretrain.py#L122

We use a BertEmb which include its own positional embedding https://github.com/LTH14/mage/blob/main/models_mage.py#L96

We use categorical sampling to generate different images. When a blank image is input to the encoder, [C0] as well as 256 [M] is input to the encoder (during generation).

We do not include a class-conditional image in this repo. The 1024+1000+1 token_emb is actually deprecated. When doing class conditional generation, we use another token_embedding which embeds the class label...

I just updated the README to include pre-trained MAGE-C checkpoints converted from JAX. We are still wrapping up the PyTorch code for the contrastive loss. We will endeavor to release...

Please refer to this [paper](https://arxiv.org/abs/2210.16870) for that comparison. We also have the comparison between MAGE and MAGE-C in Table 2 and Table 13 in our [paper](https://arxiv.org/pdf/2211.09117.pdf).

We haven't tried an encoder-only structure like BEiT. The reason why we chose the MAE-style enc-dec structure is simply that they were the sota method at that time. Also, an...

The pre-trained model we release can perform unconditional image generation on ImageNet. If you want a model for unconditional generation on your own dataset, you need to pre-train MAGE on...