ast icon indicating copy to clipboard operation
ast copied to clipboard

AST tiny and small pretrained models

Open abaronetto opened this issue 1 year ago • 1 comments

Hi Yuan, I have tested your AST mode pretrained on Audioset on my own dataset and I noticed that it achieves similar performance as EfficientNet pretrained on Audioset using psla pipeline. I was wondering if maybe a smaller AST model could achieve same performance while using less parameters (my dataset is not as big as Audioset). Did you train also tiny and small AST? Would it be possible to have AST tiny and small models pretrained on Audioset? I would like to compare them to the other models I tested (AST base, EfficientNet) and analyse the benefit of Audioset pretraining on other audio classification tasks. I will cite your work in future publications of course. Thank you in advance.

Annalisa

abaronetto avatar Jul 15 '22 05:07 abaronetto

Hi Annalisa,

Unfortunately, I don't have tiny&small ImageNet+AudioSet pretrained models. The SSAST repo has in-domian pretrained model of all sizes but it is based on a different pretraining scheme. Without AudioSet pretraining, you are free to use tiny & small models with only ImageNet pretraining.

I have tested your AST mode pretrained on Audioset on my own dataset and I noticed that it achieves similar performance as EfficientNet pretrained on Audioset using psla pipeline.

They can be similar, but I usually see a small improvement in my experiments. The AST model needs a 10X smaller learning rate than the EfficientNet model.

-Yuan

YuanGongND avatar Jul 20 '22 19:07 YuanGongND