pytorch-be-your-own-teacher icon indicating copy to clipboard operation
pytorch-be-your-own-teacher copied to clipboard

Question for your paper MSD

Open curryandsun opened this issue 4 years ago • 3 comments

Hi, Thanks for sharing this code and it's really helpful.

Recently I read your paper:"MSD: Multi-Self-Distillation Learning via Multi-classifiers within Deep Neural Networks".It's a very interesting work and the results are much better than the paper "be your own teacher" which you reimplement here.

However,after reading your paper,I could just find some slight differences between this two papers: 1.the differences of bottleneck in the model. 2.some changes of hyper-parameter.

Is there some important details that I missed?And could you please tell me about the key difference between the two papers that lead to such a significant improvement?

curryandsun avatar Nov 11 '20 13:11 curryandsun

MSD is a bi-KD or mutual learning framework while the 'be your own teacher' is a one-way knowledge distillation method

luanyunteng avatar Jan 05 '21 12:01 luanyunteng

Hi, Thanks for sharing this code and it's really helpful.

Recently I read your paper:"MSD: Multi-Self-Distillation Learning via Multi-classifiers within Deep Neural Networks".It's a very interesting work and the results are much better than the paper "be your own teacher" which you reimplement here.

However,after reading your paper,I could just find some slight differences between this two papers: 1.the differences of bottleneck in the model. 2.some changes of hyper-parameter.

Is there some important details that I missed?And could you please tell me about the key difference between the two papers that lead to such a significant improvement?

MSD is a bi-KD or mutual learning framework while the 'be your own teacher' is a one-way knowledge distillation method

luanyunteng avatar Jan 05 '21 12:01 luanyunteng

hi, Recently I read the paper "MSD" and can you provide code about this paper? thanks

yoona-unicorn avatar Feb 16 '22 09:02 yoona-unicorn