end2end_AU_speech icon indicating copy to clipboard operation
end2end_AU_speech copied to clipboard

labels.npy for arbitrary audio/video pairs

Open RaymondDixon opened this issue 5 years ago • 2 comments

Hi @haixpham, thanks for making the code available for the community. I am working on reproducing your results as part of research work. You have provided a link for AU Labels for RAVDESS dataset. How would i create/acquire Labels.npy files, if i want to do train/inference on any arbitrary video/audio pair.

RaymondDixon avatar Jul 11 '19 09:07 RaymondDixon

Thanks for your interest in my work! AU labels are extracted using my private face tracker. There is a reason it is not available:

  • The face shape model is from FaceWarehouse - they do not allow re-distribution. I was one among a few given agreement to use their face model early on. I want to replace it with a custom face model, but it will take a while. I will be happy to assist you with your research, however. Please send me the videos you need, I will extract labels and give back to you.

haixpham avatar Jul 11 '19 09:07 haixpham

Hi @haixpham thanks for providing the data and code. In the data you have provided, do the labels correspond to the standard AU[0]...AU[46] ? or are you selecting specific activation unit numbers?

meherabhi avatar Jan 22 '21 17:01 meherabhi