Transformer-SSL icon indicating copy to clipboard operation
Transformer-SSL copied to clipboard

This is an official implementation for "Self-Supervised Learning with Swin Transformers".

Results 17 Transformer-SSL issues
Sort by recently updated
recently updated
newest added

start cmd ``` imagenetpath=mypath CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 \ python -m torch.distributed.launch --nproc_per_node 8 --master_port 12345 moby_main.py \ --cfg configs/moby_swin_tiny.yaml --data-path ${imagenetpath} --batch-size 256 ``` but get the ```Gradient overflow. Skipping step, loss...

i want to use your work to perform few epochs pretrain on my dataset,which contains sevceral similar vehicle categories. So i load the imagenet-pretrained checkpoint and run another pretrain on...

Hi authors, I have pretrianed your moby_swin_tiny model using 8 Tesla V100 GPU and reproduced your results in downstream task. I get 74.394% on linear evaluation and 43.1% on COCO...

Thank you very much for this great paper. I would like to ask, will apex mixed precision training affect the accuracy of the model? I tried to install using the...

Download link of DeiT-S model: https://github.com/SwinTransformer/storage/releases/download/v1.0.3/moby_swin_t_300ep_pretrained.pth Download link of Swin-T model: https://github.com/SwinTransformer/storage/releases/download/v1.0.3/moby_deit_small_300ep_pretrained.pth ![image](https://user-images.githubusercontent.com/59405594/204264759-a7a5cf71-3112-4c2b-a567-823248132463.png) Look at the **last part** of the download link. I think the model links should be interchanged.

I am gonna train MoBY-SwinT on my custom dataset. My machine has one GPU. I tried some but failed and faced following errors. All packages are installed. * First try...

Under Transformer-SSL/data/build.py, inside the "build_transform" function, under "byol" augmentation type, the interpolation method used in RandomResizedCrop is the default which is BILINEAR, however in the BYOL paper the author used...

Traceback (most recent call last): File "moby_linear.py", line 385, in main(config) File "moby_linear.py", line 174, in main train_one_epoch(config, model, criterion, data_loader_train, optimizer, epoch, mixup_fn, lr_scheduler) File "moby_linear.py", line 199, in...

Hello dear authors, Thank you for providing your work and code. I understand from your paper that you used patch size = 4 in all your models, is there any...

Wonderful job! I recently read you code and have some questions in Swin model which is shown in swin_transformer.py. Concretely, I can't understand the calculation formula of relative_position_index and attn_mask....