TS-CAM icon indicating copy to clipboard operation
TS-CAM copied to clipboard

Question about the training process

Open Unrealluver opened this issue 2 years ago • 1 comments

Greetings! Thanks for your excellent work! When running your code, I met a problem that the performance is poor. My running command is

 bash train_val_cub.sh 3 deit small 224

and I got the log like:

{'BASIC': {'BACKUP_CODES': True,
           'BACKUP_LIST': ['lib', 'tools_cam', 'configs'],
           'DISP_FREQ': 10,
           'GPU_ID': [0],
           'NUM_WORKERS': 40,
           'ROOT_DIR': './tools_cam/..',
           'SAVE_DIR': 'ckpt/CUB/deit_tscam_small_patch16_224_CAM-NORMAL_SEED26_CAM-THR0.1_BS128_2022-03-25-01-46',
           'SEED': 26,
           'TIME': '2022-03-25-01-46'},
 'CUDNN': {'BENCHMARK': False, 'DETERMINISTIC': True, 'ENABLE': True},
 'DATA': {'CROP_SIZE': 224,
          'DATADIR': 'data/CUB_200_2011',
          'DATASET': 'CUB',
          'IMAGE_MEAN': [0.485, 0.456, 0.406],
          'IMAGE_STD': [0.229, 0.224, 0.225],
          'NUM_CLASSES': 200,
          'RESIZE_SIZE': 256,
          'SCALE_LENGTH': 15,
          'SCALE_SIZE': 196,
          'TRAIN_AUG_PATH': '',
          'VAL_PATH': ''},
 'MODEL': {'ARCH': 'deit_tscam_small_patch16_224',
           'CAM_THR': 0.1,
           'LOCALIZER_DIR': '',
           'TOP_K': 1},
 'SOLVER': {'LR_FACTOR': 0.1,
            'LR_STEPS': [30],
            'MUMENTUM': 0.9,
            'NUM_EPOCHS': 60,
            'START_LR': 0.001,
            'WEIGHT_DECAY': 0.0005},
 'TEST': {'BATCH_SIZE': 128,
          'CKPT_DIR': '',
          'SAVE_BOXED_IMAGE': False,
          'SAVE_CAMS': False,
          'TEN_CROPS': False},
 'TRAIN': {'ALPHA': 1.0,
           'BATCH_SIZE': 128,
           'BETA': 1.0,
           'IF_FIX_WEIGHT': False}}
==> Preparing data...
done!
==> Preparing networks for baseline...
Removing key head.weight from pretrained checkpoint
Removing key head.bias from pretrained checkpoint
TSCAM(
  (patch_embed): PatchEmbed(
    (proj): Conv2d(3, 384, kernel_size=(16, 16), stride=(16, 16))
  )
  (pos_drop): Dropout(p=0.0, inplace=False)
  (blocks): ModuleList(
    (0): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): Identity()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (1): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (2): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (3): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (4): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (5): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (6): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (7): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (8): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (9): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (10): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (11): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
  )
  (norm): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
  (head): Conv2d(384, 200, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (avgpool): AdaptiveAvgPool2d(output_size=1)
){'BASIC': {'BACKUP_CODES': True,
           'BACKUP_LIST': ['lib', 'tools_cam', 'configs'],
           'DISP_FREQ': 10,
           'GPU_ID': [0],
           'NUM_WORKERS': 40,
           'ROOT_DIR': './tools_cam/..',
           'SAVE_DIR': 'ckpt/CUB/deit_tscam_small_patch16_224_CAM-NORMAL_SEED26_CAM-THR0.1_BS128_2022-03-25-01-46',
           'SEED': 26,
           'TIME': '2022-03-25-01-46'},
 'CUDNN': {'BENCHMARK': False, 'DETERMINISTIC': True, 'ENABLE': True},
 'DATA': {'CROP_SIZE': 224,
          'DATADIR': 'data/CUB_200_2011',
          'DATASET': 'CUB',
          'IMAGE_MEAN': [0.485, 0.456, 0.406],
          'IMAGE_STD': [0.229, 0.224, 0.225],
          'NUM_CLASSES': 200,
          'RESIZE_SIZE': 256,
          'SCALE_LENGTH': 15,
          'SCALE_SIZE': 196,
          'TRAIN_AUG_PATH': '',
          'VAL_PATH': ''},
 'MODEL': {'ARCH': 'deit_tscam_small_patch16_224',
           'CAM_THR': 0.1,
           'LOCALIZER_DIR': '',
           'TOP_K': 1},
 'SOLVER': {'LR_FACTOR': 0.1,
            'LR_STEPS': [30],
            'MUMENTUM': 0.9,
            'NUM_EPOCHS': 60,
            'START_LR': 0.001,
            'WEIGHT_DECAY': 0.0005},
 'TEST': {'BATCH_SIZE': 128,
          'CKPT_DIR': '',
          'SAVE_BOXED_IMAGE': False,
          'SAVE_CAMS': False,
          'TEN_CROPS': False},
 'TRAIN': {'ALPHA': 1.0, 'BATCH_SIZE': 128, 'BETA': 1.0, 'IF_FIX_WEIGHT': True}}
==> Preparing data...
done!
==> Preparing networks for baseline...
Removing key head.weight from pretrained checkpoint
Removing key head.bias from pretrained checkpoint
TSCAM(
  (patch_embed): PatchEmbed(
    (proj): Conv2d(3, 384, kernel_size=(16, 16), stride=(16, 16))
  )
  (pos_drop): Dropout(p=0.0, inplace=False)
  (blocks): ModuleList(
    (0): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): Identity()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (1): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (2): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (3): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (4): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (5): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (6): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (7): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (8): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (9): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (10): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
    (11): Block(
      (norm1): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (attn): Attention(
        (qkv): Linear(in_features=384, out_features=1152, bias=True)
        (attn_drop): Dropout(p=0.0, inplace=False)
        (proj): Linear(in_features=384, out_features=384, bias=True)
        (proj_drop): Dropout(p=0.0, inplace=False)
      )
      (drop_path): DropPath()
      (norm2): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
      (mlp): Mlp(
        (fc1): Linear(in_features=384, out_features=1536, bias=True)
        (act): GELU()
        (fc2): Linear(in_features=1536, out_features=384, bias=True)
        (drop): Dropout(p=0.0, inplace=False)
      )
    )
  )
  (norm): LayerNorm((384,), eps=1e-06, elementwise_affine=True)
  (head): Conv2d(384, 200, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
  (avgpool): AdaptiveAvgPool2d(output_size=1)
)
Preparing networks done!
Train Epoch: [1][1/47],lr: 0.00005	Loss 5.3910 (5.3910)	Prec@1 0.781 (0.781)	Prec@5 3.125 (3.125)
Train Epoch: [1][11/47],lr: 0.00005	Loss 5.2794 (5.3558)	Prec@1 1.562 (0.781)	Prec@5 7.031 (2.983)
Train Epoch: [1][21/47],lr: 0.00005	Loss 5.2877 (5.3156)	Prec@1 0.781 (0.818)	Prec@5 4.688 (3.162)
Train Epoch: [1][31/47],lr: 0.00005	Loss 5.1760 (5.2851)	Prec@1 0.781 (0.882)	Prec@5 7.031 (3.931)
Train Epoch: [1][41/47],lr: 0.00005	Loss 5.1290 (5.2508)	Prec@1 6.250 (1.524)	Prec@5 12.500 (5.812)
Train Epoch: [1][47/47],lr: 0.00005	Loss 5.1377 (5.2344)	Prec@1 3.774 (1.668)	Prec@5 11.321 (6.540)
Val Epoch: [1][1/46]	Loss 4.9150 (4.9150)	
Cls@1:0.125	Cls@5:0.320
Loc@1:0.031	Loc@5:0.055	Loc_gt:0.258

Val Epoch: [1][11/46]	Loss 4.7868 (5.0117)	
Cls@1:0.065	Cls@5:0.232
Loc@1:0.011	Loc@5:0.044	Loc_gt:0.214

Val Epoch: [1][21/46]	Loss 5.0634 (5.0060)	
Cls@1:0.066	Cls@5:0.232
Loc@1:0.015	Loc@5:0.052	Loc_gt:0.217

Val Epoch: [1][31/46]	Loss 5.1113 (5.0342)	
Cls@1:0.061	Cls@5:0.206
Loc@1:0.014	Loc@5:0.046	Loc_gt:0.198

Val Epoch: [1][41/46]	Loss 5.0010 (5.0245)	
Cls@1:0.059	Cls@5:0.204
Loc@1:0.014	Loc@5:0.046	Loc_gt:0.192

Val Epoch: [1][46/46]	Loss 4.8866 (5.0296)	
Cls@1:0.059	Cls@5:0.200
Loc@1:0.013	Loc@5:0.045	Loc_gt:0.189

wrong_details:75 5454 0 6 254 5
Best GT_LOC: 0.18916120124266483
Best TOP1_LOC: 0.18916120124266483
2022-03-25-01-49
Train Epoch: [2][1/47],lr: 0.00005	Loss 5.0064 (5.0064)	Prec@1 6.250 (6.250)	Prec@5 17.969 (17.969)
Train Epoch: [2][11/47],lr: 0.00005	Loss 4.9585 (4.9966)	Prec@1 6.250 (6.818)	Prec@5 21.875 (22.656)
Train Epoch: [2][21/47],lr: 0.00005	Loss 4.9573 (4.9768)	Prec@1 8.594 (6.734)	Prec@5 28.906 (24.479)
Train Epoch: [2][31/47],lr: 0.00005	Loss 4.9050 (4.9509)	Prec@1 10.938 (7.737)	Prec@5 28.125 (25.932)
Train Epoch: [2][41/47],lr: 0.00005	Loss 4.8085 (4.9271)	Prec@1 14.844 (8.841)	Prec@5 37.500 (27.458)
Train Epoch: [2][47/47],lr: 0.00005	Loss 4.8456 (4.9160)	Prec@1 8.491 (9.059)	Prec@5 31.132 (28.195)
Val Epoch: [2][1/46]	Loss 4.5358 (4.5358)	
Cls@1:0.258	Cls@5:0.523
Loc@1:0.078	Loc@5:0.164	Loc_gt:0.344

Val Epoch: [2][11/46]	Loss 4.3821 (4.7243)	
Cls@1:0.164	Cls@5:0.431
Loc@1:0.045	Loc@5:0.109	Loc_gt:0.240

Val Epoch: [2][21/46]	Loss 4.8342 (4.6906)	
Cls@1:0.173	Cls@5:0.453
Loc@1:0.059	Loc@5:0.135	Loc_gt:0.251

Val Epoch: [2][31/46]	Loss 4.9996 (4.7545)	
Cls@1:0.153	Cls@5:0.403
Loc@1:0.050	Loc@5:0.115	Loc_gt:0.225

Val Epoch: [2][41/46]	Loss 4.8124 (4.7559)	
Cls@1:0.138	Cls@5:0.385
Loc@1:0.045	Loc@5:0.108	Loc_gt:0.217

Val Epoch: [2][46/46]	Loss 4.8159 (4.7612)	
Cls@1:0.142	Cls@5:0.391
Loc@1:0.045	Loc@5:0.108	Loc_gt:0.213

wrong_details:263 4971 0 21 536 3
Best GT_LOC: 0.2126337590610977
Best TOP1_LOC: 0.2126337590610977
2022-03-25-01-54
Train Epoch: [3][1/47],lr: 0.00005	Loss 4.7283 (4.7283)	Prec@1 21.094 (21.094)	Prec@5 46.875 (46.875)
Train Epoch: [3][11/47],lr: 0.00005	Loss 4.7234 (4.7402)	Prec@1 11.719 (15.483)	Prec@5 45.312 (43.111)
Train Epoch: [3][21/47],lr: 0.00005	Loss 4.6686 (4.7088)	Prec@1 15.625 (16.332)	Prec@5 45.312 (43.824)
Train Epoch: [3][31/47],lr: 0.00005	Loss 4.6701 (4.6906)	Prec@1 20.312 (16.608)	Prec@5 46.875 (43.800)
Train Epoch: [3][41/47],lr: 0.00005	Loss 4.5544 (4.6702)	Prec@1 26.562 (17.073)	Prec@5 50.000 (44.284)
Train Epoch: [3][47/47],lr: 0.00005	Loss 4.5622 (4.6585)	Prec@1 26.415 (17.718)	Prec@5 49.057 (44.745)
Val Epoch: [3][1/46]	Loss 4.1796 (4.1796)	
Cls@1:0.336	Cls@5:0.711
Loc@1:0.156	Loc@5:0.312	Loc_gt:0.438

Val Epoch: [3][11/46]	Loss 4.0685 (4.4652)	
Cls@1:0.263	Cls@5:0.551
Loc@1:0.078	Loc@5:0.164	Loc_gt:0.273

Val Epoch: [3][21/46]	Loss 4.6838 (4.4194)	
Cls@1:0.264	Cls@5:0.570
Loc@1:0.098	Loc@5:0.198	Loc_gt:0.294

Val Epoch: [3][31/46]	Loss 4.8199 (4.5032)	
Cls@1:0.232	Cls@5:0.515
Loc@1:0.083	Loc@5:0.167	Loc_gt:0.260

Val Epoch: [3][41/46]	Loss 4.6710 (4.5206)	
Cls@1:0.209	Cls@5:0.494
Loc@1:0.073	Loc@5:0.156	Loc_gt:0.250

Val Epoch: [3][46/46]	Loss 4.4396 (4.5273)	
Cls@1:0.213	Cls@5:0.501
Loc@1:0.072	Loc@5:0.153	Loc_gt:0.243

wrong_details:420 4557 0 45 757 15
Best GT_LOC: 0.24318260269244046
Best TOP1_LOC: 0.24318260269244046
2022-03-25-01-59
Train Epoch: [4][1/47],lr: 0.00005	Loss 4.4849 (4.4849)	Prec@1 21.875 (21.875)	Prec@5 58.594 (58.594)
Train Epoch: [4][11/47],lr: 0.00005	Loss 4.5143 (4.4929)	Prec@1 28.125 (25.284)	Prec@5 47.656 (55.185)
Train Epoch: [4][21/47],lr: 0.00005	Loss 4.3787 (4.4674)	Prec@1 22.656 (25.744)	Prec@5 56.250 (55.357)
Train Epoch: [4][31/47],lr: 0.00005	Loss 4.3940 (4.4535)	Prec@1 31.250 (25.731)	Prec@5 52.344 (54.940)
Train Epoch: [4][41/47],lr: 0.00005	Loss 4.3730 (4.4333)	Prec@1 21.875 (26.067)	Prec@5 59.375 (55.259)
Train Epoch: [4][47/47],lr: 0.00005	Loss 4.3386 (4.4240)	Prec@1 28.302 (26.376)	Prec@5 64.151 (55.672)
Val Epoch: [4][1/46]	Loss 3.8875 (3.8875)	
Cls@1:0.383	Cls@5:0.758
Loc@1:0.203	Loc@5:0.398	Loc_gt:0.484

Val Epoch: [4][11/46]	Loss 3.8129 (4.2537)	
Cls@1:0.312	Cls@5:0.580
Loc@1:0.114	Loc@5:0.204	Loc_gt:0.298

Val Epoch: [4][21/46]	Loss 4.5173 (4.1790)	
Cls@1:0.326	Cls@5:0.619
Loc@1:0.137	Loc@5:0.244	Loc_gt:0.330

Val Epoch: [4][31/46]	Loss 4.6776 (4.2892)	
Cls@1:0.285	Cls@5:0.564
Loc@1:0.115	Loc@5:0.205	Loc_gt:0.290

Val Epoch: [4][41/46]	Loss 4.4627 (4.3164)	
Cls@1:0.263	Cls@5:0.547
Loc@1:0.102	Loc@5:0.190	Loc_gt:0.277

Val Epoch: [4][46/46]	Loss 4.2653 (4.3204)	
Cls@1:0.270	Cls@5:0.557
Loc@1:0.100	Loc@5:0.186	Loc_gt:0.269

wrong_details:580 4232 0 75 889 18
Best GT_LOC: 0.26855367621677595
Best TOP1_LOC: 0.26855367621677595
2022-03-25-02-01
Train Epoch: [5][1/47],lr: 0.00005	Loss 4.3349 (4.3349)	Prec@1 32.812 (32.812)	Prec@5 57.031 (57.031)
Train Epoch: [5][11/47],lr: 0.00005	Loss 4.2210 (4.2754)	Prec@1 31.250 (33.239)	Prec@5 62.500 (62.713)
Train Epoch: [5][21/47],lr: 0.00005	Loss 4.2603 (4.2594)	Prec@1 31.250 (32.626)	Prec@5 57.812 (61.793)
Train Epoch: [5][31/47],lr: 0.00005	Loss 4.2397 (4.2502)	Prec@1 29.688 (31.678)	Prec@5 62.500 (61.164)
Train Epoch: [5][41/47],lr: 0.00005	Loss 4.2377 (4.2285)	Prec@1 26.562 (31.155)	Prec@5 60.156 (61.300)
Train Epoch: [5][47/47],lr: 0.00005	Loss 4.1144 (4.2206)	Prec@1 34.906 (31.365)	Prec@5 59.434 (61.328)
Val Epoch: [5][1/46]	Loss 3.6491 (3.6491)	
Cls@1:0.398	Cls@5:0.789
Loc@1:0.203	Loc@5:0.445	Loc_gt:0.516

Val Epoch: [5][11/46]	Loss 3.5341 (4.0492)	
Cls@1:0.343	Cls@5:0.620
Loc@1:0.140	Loc@5:0.247	Loc_gt:0.333

Val Epoch: [5][21/46]	Loss 4.3516 (3.9736)	
Cls@1:0.353	Cls@5:0.653
Loc@1:0.156	Loc@5:0.279	Loc_gt:0.361

Val Epoch: [5][31/46]	Loss 4.5599 (4.1005)	
Cls@1:0.306	Cls@5:0.605
Loc@1:0.132	Loc@5:0.238	Loc_gt:0.318

Val Epoch: [5][41/46]	Loss 4.3230 (4.1360)	
Cls@1:0.286	Cls@5:0.593
Loc@1:0.121	Loc@5:0.225	Loc_gt:0.306

Val Epoch: [5][46/46]	Loss 4.1012 (4.1362)	
Cls@1:0.295	Cls@5:0.602
Loc@1:0.119	Loc@5:0.220	Loc_gt:0.297

wrong_details:688 4082 0 88 912 24
Best GT_LOC: 0.2965136347946151
Best TOP1_LOC: 0.2965136347946151
2022-03-25-02-02
Train Epoch: [6][1/47],lr: 0.00005	Loss 4.1231 (4.1231)	Prec@1 30.469 (30.469)	Prec@5 66.406 (66.406)
Train Epoch: [6][11/47],lr: 0.00005	Loss 4.0252 (4.0962)	Prec@1 37.500 (35.085)	Prec@5 70.312 (67.756)
Train Epoch: [6][21/47],lr: 0.00005	Loss 3.9509 (4.0630)	Prec@1 40.625 (36.533)	Prec@5 67.969 (67.671)
Train Epoch: [6][31/47],lr: 0.00005	Loss 3.8919 (4.0431)	Prec@1 45.312 (36.215)	Prec@5 64.844 (67.137)
Train Epoch: [6][41/47],lr: 0.00005	Loss 3.9957 (4.0417)	Prec@1 40.625 (36.128)	Prec@5 70.312 (66.749)
Train Epoch: [6][47/47],lr: 0.00005	Loss 3.9811 (4.0315)	Prec@1 41.509 (36.303)	Prec@5 64.151 (67.050)
Val Epoch: [6][1/46]	Loss 3.4300 (3.4300)	
Cls@1:0.438	Cls@5:0.781
Loc@1:0.219	Loc@5:0.453	Loc_gt:0.531

Val Epoch: [6][11/46]	Loss 3.3890 (3.8868)	
Cls@1:0.357	Cls@5:0.643
Loc@1:0.145	Loc@5:0.259	Loc_gt:0.343

Val Epoch: [6][21/46]	Loss 4.1725 (3.7921)	
Cls@1:0.377	Cls@5:0.680
Loc@1:0.170	Loc@5:0.299	Loc_gt:0.379

Val Epoch: [6][31/46]	Loss 4.4162 (3.9271)	
Cls@1:0.331	Cls@5:0.634
Loc@1:0.146	Loc@5:0.259	Loc_gt:0.336

Val Epoch: [6][41/46]	Loss 4.2253 (3.9698)	
Cls@1:0.313	Cls@5:0.623
Loc@1:0.136	Loc@5:0.245	Loc_gt:0.325

Val Epoch: [6][46/46]	Loss 3.9466 (3.9713)	
Cls@1:0.321	Cls@5:0.632
Loc@1:0.134	Loc@5:0.239	Loc_gt:0.314

wrong_details:776 3935 0 118 940 25
Best GT_LOC: 0.3139454608215395
Best TOP1_LOC: 0.3139454608215395
Preparing networks done!
Train Epoch: [1][1/47],lr: 0.00005	Loss 5.3910 (5.3910)	Prec@1 0.781 (0.781)	Prec@5 3.125 (3.125)
Train Epoch: [1][11/47],lr: 0.00005	Loss 5.2530 (5.3511)	Prec@1 1.562 (0.923)	Prec@5 3.906 (3.125)
Train Epoch: [1][21/47],lr: 0.00005	Loss 5.2631 (5.3216)	Prec@1 0.781 (0.781)	Prec@5 3.906 (3.646)
Train Epoch: [1][31/47],lr: 0.00005	Loss 5.1785 (5.2905)	Prec@1 0.781 (0.907)	Prec@5 3.125 (4.461)
Train Epoch: [1][41/47],lr: 0.00005	Loss 5.1472 (5.2599)	Prec@1 1.562 (0.953)	Prec@5 7.031 (4.668)
Train Epoch: [1][47/47],lr: 0.00005	Loss 5.1461 (5.2453)	Prec@1 1.887 (1.001)	Prec@5 12.264 (4.805)
Val Epoch: [1][1/46]	Loss 4.8300 (4.8300)	
Cls@1:0.000	Cls@5:0.094
Loc@1:0.000	Loc@5:0.023	Loc_gt:0.312

Val Epoch: [1][11/46]	Loss 4.7840 (5.0671)	
Cls@1:0.010	Cls@5:0.077
Loc@1:0.002	Loc@5:0.014	Loc_gt:0.224

Val Epoch: [1][21/46]	Loss 5.3839 (5.0786)	
Cls@1:0.010	Cls@5:0.070
Loc@1:0.002	Loc@5:0.016	Loc_gt:0.218

Val Epoch: [1][31/46]	Loss 5.3107 (5.1220)	
Cls@1:0.010	Cls@5:0.061
Loc@1:0.003	Loc@5:0.014	Loc_gt:0.199

Val Epoch: [1][41/46]	Loss 4.7929 (5.0675)	
Cls@1:0.016	Cls@5:0.069
Loc@1:0.005	Loc@5:0.016	Loc_gt:0.195

Val Epoch: [1][46/46]	Loss 5.0628 (5.0798)	
Cls@1:0.016	Cls@5:0.066
Loc@1:0.005	Loc@5:0.015	Loc_gt:0.192

wrong_details:27 5704 0 3 58 2
Best GT_LOC: 0.1924404556437694
Best TOP1_LOC: 0.1924404556437694
2022-03-25-01-49
Train Epoch: [2][1/47],lr: 0.00005	Loss 5.0344 (5.0344)	Prec@1 2.344 (2.344)	Prec@5 7.812 (7.812)
Train Epoch: [2][11/47],lr: 0.00005	Loss 4.9748 (5.0317)	Prec@1 0.781 (1.634)	Prec@5 9.375 (7.812)
Train Epoch: [2][21/47],lr: 0.00005	Loss 4.8753 (4.9973)	Prec@1 3.906 (2.083)	Prec@5 10.938 (8.557)
Train Epoch: [2][31/47],lr: 0.00005	Loss 4.8447 (4.9587)	Prec@1 1.562 (2.419)	Prec@5 9.375 (9.199)
Train Epoch: [2][41/47],lr: 0.00005	Loss 4.9252 (4.9204)	Prec@1 0.781 (2.591)	Prec@5 10.938 (10.061)
Train Epoch: [2][47/47],lr: 0.00005	Loss 4.7829 (4.8979)	Prec@1 4.717 (2.669)	Prec@5 16.981 (10.777)
Val Epoch: [2][1/46]	Loss 4.8567 (4.8567)	
Cls@1:0.008	Cls@5:0.109
Loc@1:0.008	Loc@5:0.078	Loc_gt:0.352

Val Epoch: [2][11/46]	Loss 4.4700 (4.4981)	
Cls@1:0.062	Cls@5:0.232
Loc@1:0.018	Loc@5:0.061	Loc_gt:0.241

Val Epoch: [2][21/46]	Loss 4.9166 (4.5804)	
Cls@1:0.057	Cls@5:0.199
Loc@1:0.015	Loc@5:0.049	Loc_gt:0.231

Val Epoch: [2][31/46]	Loss 4.8543 (4.6311)	
Cls@1:0.055	Cls@5:0.184
Loc@1:0.013	Loc@5:0.044	Loc_gt:0.230

Val Epoch: [2][41/46]	Loss 4.4745 (4.5779)	
Cls@1:0.052	Cls@5:0.182
Loc@1:0.013	Loc@5:0.046	Loc_gt:0.217

Val Epoch: [2][46/46]	Loss 3.8181 (4.5981)	
Cls@1:0.051	Cls@5:0.177
Loc@1:0.012	Loc@5:0.044	Loc_gt:0.215

wrong_details:71 5500 0 116 98 9
Best GT_LOC: 0.21453227476700035
Best TOP1_LOC: 0.21453227476700035
2022-03-25-01-54
Train Epoch: [3][1/47],lr: 0.00005	Loss 4.4202 (4.4202)	Prec@1 7.031 (7.031)	Prec@5 21.875 (21.875)
Train Epoch: [3][11/47],lr: 0.00005	Loss 4.2909 (4.5166)	Prec@1 8.594 (5.114)	Prec@5 26.562 (19.389)
Train Epoch: [3][21/47],lr: 0.00005	Loss 4.5282 (4.5042)	Prec@1 1.562 (5.320)	Prec@5 11.719 (19.085)
Train Epoch: [3][31/47],lr: 0.00005	Loss 4.3235 (4.4662)	Prec@1 9.375 (5.872)	Prec@5 21.875 (19.481)
Train Epoch: [3][41/47],lr: 0.00005	Loss 4.2087 (4.4195)	Prec@1 8.594 (6.250)	Prec@5 31.250 (21.418)
Train Epoch: [3][47/47],lr: 0.00005	Loss 4.1792 (4.3886)	Prec@1 7.547 (6.390)	Prec@5 22.642 (21.955)
Val Epoch: [3][1/46]	Loss 3.8698 (3.8698)	
Cls@1:0.172	Cls@5:0.289
Loc@1:0.055	Loc@5:0.102	Loc_gt:0.375

Val Epoch: [3][11/46]	Loss 4.1646 (3.9793)	
Cls@1:0.118	Cls@5:0.327
Loc@1:0.029	Loc@5:0.101	Loc_gt:0.257

Val Epoch: [3][21/46]	Loss 4.8005 (4.0862)	
Cls@1:0.106	Cls@5:0.299
Loc@1:0.022	Loc@5:0.076	Loc_gt:0.228

Val Epoch: [3][31/46]	Loss 4.6957 (4.3865)	
Cls@1:0.085	Cls@5:0.241
Loc@1:0.017	Loc@5:0.061	Loc_gt:0.225

Val Epoch: [3][41/46]	Loss 4.3491 (4.2824)	
Cls@1:0.082	Cls@5:0.253
Loc@1:0.017	Loc@5:0.062	Loc_gt:0.210

Val Epoch: [3][46/46]	Loss 3.9240 (4.3113)	
Cls@1:0.079	Cls@5:0.246
Loc@1:0.016	Loc@5:0.057	Loc_gt:0.203

wrong_details:90 5334 0 314 42 14
Best GT_LOC: 0.21453227476700035
Best TOP1_LOC: 0.21453227476700035
2022-03-25-02-00
Train Epoch: [4][1/47],lr: 0.00005	Loss 3.9914 (3.9914)	Prec@1 12.500 (12.500)	Prec@5 34.375 (34.375)
Train Epoch: [4][11/47],lr: 0.00005	Loss 4.1547 (4.0524)	Prec@1 10.156 (11.364)	Prec@5 29.688 (30.753)
Train Epoch: [4][21/47],lr: 0.00005	Loss 4.3899 (4.0732)	Prec@1 7.031 (10.528)	Prec@5 19.531 (30.432)
Train Epoch: [4][31/47],lr: 0.00005	Loss 3.7553 (4.0368)	Prec@1 11.719 (10.459)	Prec@5 38.281 (31.401)
Train Epoch: [4][41/47],lr: 0.00005	Loss 3.9345 (4.0095)	Prec@1 8.594 (10.690)	Prec@5 30.469 (31.879)
Train Epoch: [4][47/47],lr: 0.00005	Loss 3.7705 (3.9801)	Prec@1 16.981 (11.195)	Prec@5 38.679 (32.733)
Val Epoch: [4][1/46]	Loss 4.2226 (4.2226)	
Cls@1:0.141	Cls@5:0.344
Loc@1:0.086	Loc@5:0.195	Loc_gt:0.461

Val Epoch: [4][11/46]	Loss 3.6754 (3.8854)	
Cls@1:0.148	Cls@5:0.378
Loc@1:0.048	Loc@5:0.126	Loc_gt:0.303

Val Epoch: [4][21/46]	Loss 4.4785 (3.9334)	
Cls@1:0.134	Cls@5:0.353
Loc@1:0.036	Loc@5:0.102	Loc_gt:0.283

Val Epoch: [4][31/46]	Loss 4.5506 (3.9342)	
Cls@1:0.123	Cls@5:0.345
Loc@1:0.035	Loc@5:0.108	Loc_gt:0.294

Val Epoch: [4][41/46]	Loss 3.8479 (3.8611)	
Cls@1:0.121	Cls@5:0.354
Loc@1:0.030	Loc@5:0.100	Loc_gt:0.267

Val Epoch: [4][46/46]	Loss 3.7787 (3.8731)	
Cls@1:0.118	Cls@5:0.345
Loc@1:0.028	Loc@5:0.094	Loc_gt:0.265

wrong_details:164 5109 0 465 41 15
Best GT_LOC: 0.26458405246807043
Best TOP1_LOC: 0.26458405246807043
2022-03-25-02-02
Train Epoch: [5][1/47],lr: 0.00005	Loss 3.6695 (3.6695)	Prec@1 11.719 (11.719)	Prec@5 38.281 (38.281)
Train Epoch: [5][11/47],lr: 0.00005	Loss 3.5674 (3.5657)	Prec@1 19.531 (16.548)	Prec@5 40.625 (42.827)
Train Epoch: [5][21/47],lr: 0.00005	Loss 3.5742 (3.5837)	Prec@1 13.281 (16.481)	Prec@5 40.625 (43.043)
Train Epoch: [5][31/47],lr: 0.00005	Loss 3.5905 (3.5552)	Prec@1 19.531 (17.339)	Prec@5 44.531 (43.901)
Train Epoch: [5][41/47],lr: 0.00005	Loss 3.4860 (3.5470)	Prec@1 22.656 (17.492)	Prec@5 46.875 (44.284)
Train Epoch: [5][47/47],lr: 0.00005	Loss 3.6738 (3.5731)	Prec@1 13.208 (17.017)	Prec@5 38.679 (43.694)
Val Epoch: [5][1/46]	Loss 3.6581 (3.6581)	
Cls@1:0.133	Cls@5:0.398
Loc@1:0.039	Loc@5:0.164	Loc_gt:0.328

Val Epoch: [5][11/46]	Loss 3.4856 (3.7789)	
Cls@1:0.167	Cls@5:0.392
Loc@1:0.037	Loc@5:0.097	Loc_gt:0.261

Val Epoch: [5][21/46]	Loss 3.7290 (3.6464)	
Cls@1:0.188	Cls@5:0.424
Loc@1:0.036	Loc@5:0.093	Loc_gt:0.240

Val Epoch: [5][31/46]	Loss 4.2331 (3.7345)	
Cls@1:0.165	Cls@5:0.397
Loc@1:0.036	Loc@5:0.094	Loc_gt:0.242

Val Epoch: [5][41/46]	Loss 4.0443 (3.7508)	
Cls@1:0.156	Cls@5:0.393
Loc@1:0.034	Loc@5:0.089	Loc_gt:0.227

Val Epoch: [5][46/46]	Loss 3.5688 (3.7679)	
Cls@1:0.150	Cls@5:0.389
Loc@1:0.032	Loc@5:0.086	Loc_gt:0.226

wrong_details:185 4926 0 622 34 27
Best GT_LOC: 0.26458405246807043
Best TOP1_LOC: 0.26458405246807043

Looking for your help!

Unrealluver avatar Apr 20 '22 07:04 Unrealluver

If training model under the same setting as paper, the result should not differ a lot. Could you try running the code on one or two GPUs?

vasgaowei avatar Apr 21 '22 11:04 vasgaowei