autodeeplab icon indicating copy to clipboard operation
autodeeplab copied to clipboard

Slow Training, any Advice?

Open albert-ba opened this issue 5 years ago • 7 comments

Hi, I'm using this command line: CUDA_VISIBLE_DEVICES=0,1,2,3 python3 train_autodeeplab.py --dataset coco --filter_multiplier 4 --resize 358 --crop_size 224 --batch-size 24

I'm trying to understand why is it so slow, I'm getting an epoch per day, in the paper they were talking about 3 days for 40 epoch. My question is this a known issue, or that I'm missing something? I'll be happy for any advice

Thanks!

albert-ba avatar Aug 31 '19 07:08 albert-ba

Hi @albert-ba, for now I suggest training on 1 GPU with 2 batch size. It will take you a couple hours per epoch. It’s still not fast enough but we’re working on it.

Once we’ve sufficiently sped up the code on a single GPU, we’ll start looking into speeding up multi-GPU.

All the best!

NoamRosenberg avatar Sep 02 '19 17:09 NoamRosenberg

Hi it doesn't seem to help. image

98 hours to finish the epoch..

I know that this repo in still under optimization, I have no complains at all you are doing wonderfull job, but if think about something susspicous that I'm doing wrong please share. maybe related to those warnings? image

albert-ba avatar Sep 03 '19 06:09 albert-ba

what GPUs are you running on? these warnings are known and can be ignored for now, i don't think they're related to your issue at all.

I see that you used the coco dataset. we usually conduct all our experiments with cityscapes. could you try to train with cityscapes with the default args and see what you get?

iariav avatar Sep 03 '19 10:09 iariav

hi,I met the same question with you. I'm getting an epoch per day. I start training with this command line. CUDA_VISIBLE_DEVICES=0 python train_autodeeplab.py --dataset cityscapes

I use a single GPU quadro gv100, training with cityscapes. Batch size is 2. Other I am configured by default. Have you solved this problem?@albert-ba Do you have any advice?@NoamRosenberg @iariav

fanrupin avatar Nov 13 '19 15:11 fanrupin

@fanrupin currently the multi-gpu doesn't always work faster. I have to look into this further but I don't have time now. If you have time to experiment and find the problem I would love you to make a pull request and become a contributer.

NoamRosenberg avatar Nov 19 '19 10:11 NoamRosenberg

Hi, I'm using this command line: CUDA_VISIBLE_DEVICES=0,1,2,3 python3 train_autodeeplab.py --dataset coco --filter_multiplier 4 --resize 358 --crop_size 224 --batch-size 24

I'm trying to understand why is it so slow, I'm getting an epoch per day, in the paper they were talking about 3 days for 40 epoch. My question is this a known issue, or that I'm missing something? I'll be happy for any advice

Thanks!

Hello, I also want to use this code on COCO dataset. I adopted the command line just like yours: CUDA_VISIBLE_DEVICES=0,1 python3 train_autodeeplab.py --dataset coco --filter_multiplier 4 --resize 358 --crop_size 224 --batch-size 24 ----gpu-ids 0,1

But I got an error: Untitled I wanna know whether you have met this error? If yes, could you please tell my how to solve it?

Thanks a lot!

MoExplorer avatar Sep 27 '20 08:09 MoExplorer

Hi, I'm using this command line: CUDA_VISIBLE_DEVICES=0,1,2,3 python3 train_autodeeplab.py --dataset coco --filter_multiplier 4 --resize 358 --crop_size 224 --batch-size 24 I'm trying to understand why is it so slow, I'm getting an epoch per day, in the paper they were talking about 3 days for 40 epoch. My question is this a known issue, or that I'm missing something? I'll be happy for any advice Thanks!

Hello, I also want to use this code on COCO dataset. I adopted the command line just like yours: CUDA_VISIBLE_DEVICES=0,1 python3 train_autodeeplab.py --dataset coco --filter_multiplier 4 --resize 358 --crop_size 224 --batch-size 24 ----gpu-ids 0,1

But I got an error: Untitled I wanna know whether you have met this error? If yes, could you please tell my how to solve it?

Thanks a lot!

you need to change the crop size

ZhiboRao avatar Jan 05 '21 08:01 ZhiboRao