end2end_AU_speech
end2end_AU_speech copied to clipboard
labels.npy for arbitrary audio/video pairs
Hi @haixpham, thanks for making the code available for the community. I am working on reproducing your results as part of research work. You have provided a link for AU Labels for RAVDESS dataset. How would i create/acquire Labels.npy files, if i want to do train/inference on any arbitrary video/audio pair.
Thanks for your interest in my work! AU labels are extracted using my private face tracker. There is a reason it is not available:
- The face shape model is from FaceWarehouse - they do not allow re-distribution. I was one among a few given agreement to use their face model early on. I want to replace it with a custom face model, but it will take a while. I will be happy to assist you with your research, however. Please send me the videos you need, I will extract labels and give back to you.
Hi @haixpham thanks for providing the data and code. In the data you have provided, do the labels correspond to the standard AU[0]...AU[46] ? or are you selecting specific activation unit numbers?