Skip to content

SoyeonHH/CDaT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Python 3.7

CDaT

Pytorch implementation for Cross-Modal Dynamic Transfer Learning for Multi-Modal Emotion Recognition

Paper

Cross-Modal Dynamic Transfer Learning for Multi-Modal Emotion Recognition (Accepted at IEEE Access)

Overview

Overall Architecture for CDaT

We proposed a representation learning method called Cross-Modal Dynamic Transfer Learning (CDaT), which dynamically filters the low-confident modality and complements it with the high-confident modality using uni-modal masking and cross-modal representation transfer learning. We train an auxiliary network that learns model confidence scores to determine which modality is low-confident and how much the transfer should occur from other modalities. Furthermore, it can be used with any fusion model in a model-agnostic way because it leverages transfer between low-level uni-modal information via probabilistic knowledge transfer loss.

Usage

Datasets

get started

  1. the first step is clone this repo
git clone [email protected]:SoyeonHH/CDaT.git
  1. Set up the environment (need conda prerequisities) from the environment.yml file
conda env create -f environment.yml
bash init.sh
  1. Download the dataset and put it in the data folder & Modify the path in train.py
mosei_data_dir = 'YOUR_PATH'
iemocap_data_dir = 'YOUR_PATH'
  1. Train the model
bash train.sh
  1. Test the model
bash inference.sh

Note that modify the model path and data path

Acknowledgement

Some portion of the code were adapted from the TAILOR repo. We thank the authors for their great work.

About

2024 IEEE Access Published Paper - Cross-Modal Dynamic Transfer Learning for Multi-Modal Emotion Recognition

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published