Skip to content

Gao-xiyuan/ARM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

34 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ARM: Asymmetric Reinforcing against Multimodal Representation Bias

This repository contains the official implementation of our paper:
"Asymmetric Reinforcing against Multimodal Representation Bias"
Developed based on the MMPretrain framework.

image

🔍 Introduction

The strength of multimodal learning lies in its ability to integrate information from various sources, providing rich and comprehensive insights. However, in real-world scenarios, multi-modal systems often face the challenge of dynamic modality contributions, the dominance of different modalities may change with the environments, leading to suboptimal performance in multimodal learning. Current methods mainly enhance weak modalities to balance multimodal representation bias, which inevitably optimizes from a partialmodality perspective, easily leading to performance descending for dominant modalities. To address this problem, we propose an Asymmetric Reinforcing method against Multimodal representation bias (ARM). Our ARM dynamically reinforces the weak modalities while maintaining the ability to represent dominant modalities through conditional mutual information. Moreover, we provide an in-depth analysis that optimizing certain modalities could cause information loss and prevent leveraging the full advantages of multimodal data. By exploring the dominance and narrowing the contribution gaps between modalities, we have significantly improved the performance of multimodal learning, making notable progress in mitigating imbalanced multimodal learning.

📂 Project Structure

arm/
├── configs/              # Configuration files for training
├── mmpretrain/           # Core implementation of ARM
├── tools/                # Training and evaluation scripts
├── work_dirs/            # Results and training logs
└── README.md

🚀 Getting Started

1. Environment Setup

Install MMPretrain and dependencies:

Ensure mmpretrain v1.2.0 is installed and available in your environment.

2. Dataset Preparation

Please refer to OGM-GE.

3. Training

Run the training using the provided config:

python tools/train.py configs/arm/resnet18_ks.py

4. Evaluation

python tools/test.py configs/arm/resnet18_ks.py checkpoints/best.pth

📦 Pretrained Models and Training Logs

We provide pretrained models and training logs for reproducibility:

Dataset Backbone Config Checkpoint Training Log
Kinetics-Sounds ResNet18 resnet18_ks.py Download Log

📄 Citation

If you find our work useful, please consider citing:

@inproceedings{gao2025asymmetric,
  title={Asymmetric Reinforcing against Multi-modal Representation Bias},
  author={Gao, Xiyuan and Cao, Bing and Zhu, Pengfei and Wang, Nannan and Hu, Qinghua},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={39},
  number={16},
  pages={16754-16762},
  year={2025}
}

🤝 Acknowledgements

This project is built upon the excellent MMPretrain framework by OpenMMLab and OGM-GE by GeWu-Lab.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published