Feature extraction for MELD dataset #57
Unanswered
siddhantpathakk
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am using the MELD dataset and the MELD features from the pickle file under https://github.com/declare-lab/conv-emotion/blob/master/DialogueRNN/DialogueRNN_features.zip
In the original paper for DialogueRNN, it mentions about using openSMILE toolkit to retrieve the audio and text features, as (300,0) or (600,0( vectors for each utterance.
Could you explain the pre-processing done on the raw mp4 files to get these vectors using the toolkit? I wish to utilize and make an inference system, I have the mp4 files and I wish to use the openSMILE toolkit to get the embeddings..
Beta Was this translation helpful? Give feedback.
All reactions