declare-lab/MELD

Using pre-trained models with our own audio files.

vsakkas96 opened this issue · 2 comments

Hi, thanks for making this open source project.

I'm trying to use the pre-trained models you provide on my own audio files in order to extract the emotion and sentiment labels, but the baseline.py does not seem to provide a way to use my own files.

Moreover, the baseline.py file loads .pkl instead of .wav or .mp4 files. How would I go about using my own files an generating similar .pkl files in order to be used with the pre-trained models?

Thanks.

I am also interested in doing this. Would you be willing to share the code that extracts the text embeddings (Glove + dimensional reduction) & audio embeddings (OpenSmile + dimensional reduction)? Thanks!

Hi,

We encourage users to explore novel methods of feature selection to fit our framework. To follow our basic process mentioned in the paper, users can utilize open-sourced scripts like scikit for feature selection and opensmile for extraction of audio functionals.

Thanks!