D-TDNN icon indicating copy to clipboard operation
D-TDNN copied to clipboard

questions about model training

Open forwiat opened this issue 4 years ago • 4 comments

hello, yuyq96, Thank you so much for the great work you've shared. I learned that D-TDNNSS mini-batch setting 128 from D-TDNN paper. But this model is too large to train on single gpu. Could you tell me how you train it? Using nn.Parallel or DDP? Looking forward to you reply

forwiat avatar Dec 22 '20 06:12 forwiat

Actually, it should fit in a GPU with 12GB RAM when memory_efficient set to True, which is the default. If not, you can try:

  • Using nn.DataParallel to train the model on two cards, and the mini-batch size on each card is 64, which should also be large enough for the batch normalization.
  • Slightly decreasing the mini-batch size, e.g. to 100, then see if it fits in your card and how much RAM it consumes.

yuyq96 avatar Dec 22 '20 07:12 yuyq96

Ok, I will try it. Thanks a lot

forwiat avatar Dec 22 '20 09:12 forwiat

@yuyq96 Hi, do you have a training script for this model?

shgidi avatar May 02 '23 14:05 shgidi

@shgidi The original training script for D-TDNN or CAM are not open source due to the lack of company procedure. However, now you can use 3D-Speaker to train D-TDNN, CAM and CAM++. This project is lead by my former colleagues at Alibaba DAMO, and it is similar to the original training script.

yuyq96 avatar May 04 '23 08:05 yuyq96