FT-w2v2-ser icon indicating copy to clipboard operation
FT-w2v2-ser copied to clipboard

Official implementation for the paper Exploring Wav2vec 2.0 fine-tuning for improved speech emotion recognition

Results 8 FT-w2v2-ser issues
Sort by recently updated
recently updated
newest added

Hello, I want to know about the values of NUM_EXP because I cannot obtain the correct confusion matrix when performing each fine-tuning method. The results are : Testing: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1085/1085...

1. What's the rationale behind making the default batch size 64 for the pre-training, continued pre-training, and fine-tuning loops? Others have mentioned that they had to reduce the batch size...

Got the following when running run_downstream_custom_multiple_fold.py RuntimeError: CUDA out of memory. Tried to allocate 730.00 MiB (GPU 0; 23.70 GiB total capacity; 21.65 GiB already allocated; 426.81 MiB free; 21.81...

cp: Unable to get'Dataset/IEMOCAP/labels_sess/label_{SESSION_TO_TEST}.json' s File status(stat): There is no file or directory cp: 无法获取'Dataset/IEMOCAP/labels_sess/label_{SESSION_TO_TEST}.json' 的文件状态(stat): 没有那个文件或目录

Thank you for your share, I am new to SER, and I am learning this code recently. I would like to ask how to test it. Is there a reference...

We cannot reproduce the results about TAPT, and our pretraining loss is ''nan'' when running ''FT-w2v2-ser-main\run_baseline_continueFT.py''. Can you help us solve this issue?

1. I download the MELD Datasets from [https://affective-meld.github.io/](https://affective-meld.github.io/) 2. Then I do some lable mapping which map the ‘joy’ to 'happy' ,'sadness' to 'sad' by this script: ``` if x...