Skip to content

Multi-modal Collaborative Optimization and Expansion Network for Event-assisted Single-eye Expression Recognition

Notifications You must be signed in to change notification settings

hrdhrd/MCO-E-Net

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Multi-modal Collaborative Optimization and Expansion Network for Event-assisted Single-eye Expression Recognition

Abstract

In this paper, we proposed a Multi-modal Collaborative Optimization and Expansion Network (MCO-E Net), to use event modalities to resist challenges such as low light, high exposure, and high dynamic range in single-eye expression recognition tasks. The MCO-E Net introduces two innovative designs: Multi-modal Collaborative Optimization Mamba (MCO-Mamba) and Heterogeneous Collaborative and Expansion Mixture-of-Experts (HCE-MoE). MCO-Mamba, building upon Mamba, leverages dual-modal information to jointly optimize the model, facilitating collaborative interaction and fusion of modal semantics. This approach encourages the model to balance the learning of both modalities and harness their respective strengths. HCE-MoE, on the other hand, employs a dynamic routing mechanism to distribute structurally varied experts (deep, attention, and focal), fostering collaborative learning of complementary semantics. This heterogeneous architecture systematically integrates diverse feature extraction paradigms to comprehensively capture expression semantics.

Overview

Start

git clone https://github.com/hrdhrd/MCO-E-Net
cd MCO-E-Net
unzip code.zip

Datasets

Apply for SEE dataset.

Modify the dataset path in the opt.py file: "--event_video_path" and "--frame_video_path"

Train

To train the model, first:

pip install -r requirements.txt
mkdir your_save_path
cd code

Then, replace the code in ./mamba_ssm/ops/selective_scan_interface.py with the code in BC.py

Final,

CUDA_VISIBLE_DEVICES=0  python train.py   --result_path  your_save_path   --inference  --tensorboard --sample_duration 4    --sample_t_stride 4  --inference_sample_duration 4

Evaluation

To evaluate the model, run:

CUDA_VISIBLE_DEVICES=0  python evaluation.py   --result_path  your_save_path   --inference  --tensorboard --sample_duration 4    --sample_t_stride 4  --inference_sample_duration 4 

Accuracy

Calculate UAR, WAR:

python read_20json_result.py

About

Multi-modal Collaborative Optimization and Expansion Network for Event-assisted Single-eye Expression Recognition

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages