Transformer-SSL
Transformer-SSL copied to clipboard
This is an official implementation for "Self-Supervised Learning with Swin Transformers".
start cmd ``` imagenetpath=mypath CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 \ python -m torch.distributed.launch --nproc_per_node 8 --master_port 12345 moby_main.py \ --cfg configs/moby_swin_tiny.yaml --data-path ${imagenetpath} --batch-size 256 ``` but get the ```Gradient overflow. Skipping step, loss...
i want to use your work to perform few epochs pretrain on my dataset,which contains sevceral similar vehicle categories. So i load the imagenet-pretrained checkpoint and run another pretrain on...
Hi authors, I have pretrianed your moby_swin_tiny model using 8 Tesla V100 GPU and reproduced your results in downstream task. I get 74.394% on linear evaluation and 43.1% on COCO...
Thank you very much for this great paper. I would like to ask, will apex mixed precision training affect the accuracy of the model? I tried to install using the...
Download link of DeiT-S model: https://github.com/SwinTransformer/storage/releases/download/v1.0.3/moby_swin_t_300ep_pretrained.pth Download link of Swin-T model: https://github.com/SwinTransformer/storage/releases/download/v1.0.3/moby_deit_small_300ep_pretrained.pth  Look at the **last part** of the download link. I think the model links should be interchanged.
I am gonna train MoBY-SwinT on my custom dataset. My machine has one GPU. I tried some but failed and faced following errors. All packages are installed. * First try...
Under Transformer-SSL/data/build.py, inside the "build_transform" function, under "byol" augmentation type, the interpolation method used in RandomResizedCrop is the default which is BILINEAR, however in the BYOL paper the author used...
Traceback (most recent call last): File "moby_linear.py", line 385, in main(config) File "moby_linear.py", line 174, in main train_one_epoch(config, model, criterion, data_loader_train, optimizer, epoch, mixup_fn, lr_scheduler) File "moby_linear.py", line 199, in...
Hello dear authors, Thank you for providing your work and code. I understand from your paper that you used patch size = 4 in all your models, is there any...
Wonderful job! I recently read you code and have some questions in Swin model which is shown in swin_transformer.py. Concretely, I can't understand the calculation formula of relative_position_index and attn_mask....