Skip to content

injadlu/DAMA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 

Repository files navigation

DAMA: Data- and Model-aware Alignment of Multi-modal LLMs

Implementation of the ICML 2025 paper: Data- and Model-aware Alignment of Multi-modal LLMs.

Introduction

DAMA: \textbf{Da}ta- and \textbf{M}odel-\textbf{a}ware DPO is a new method to dynamically adjust the preference optimization process of Multi-modal LLMs, and it consists of the following two aspects: (1) a data-aware strategy that incorporates data hardness, and (2) a model-aware strategy that integrates real-time model responses. By combining the two strategies, DAMA enables the model to effectively adapt to data with varying levels of hardness. Extensive experiments on five benchmarks demonstrate that DAMA not only significantly enhances the trustworthiness, but also improves the effectiveness over general tasks. For instance, on the Object HalBench, our DAMA-7B reduces response-level and mentioned-level hallucination by 90.0% and 95.3%, respectively. \vspace{1em}

Data-aware strategy


This strategy incorporates data hardness to modulate optimization strength based on the data hardness.

Model-aware strategy


This strategy dynamically integrates real-time model responses to reflect its current state.

Get Started

Dataset: We employ the LLaVA-v1.5-7B preference data from RLAIF-V Dataset
Code: DAMA code is built upon LLaVA-v1.5
Download our train_dpo.py file, and insert into LLaVA codebase.
This setup allows for easy integration and execution of our DPO training pipeline.

contact:

Any problems please contact me at [email protected]

References

@misc{lu2025damadatamodelawarealignment,
      title={DAMA: Data- and Model-aware Alignment of Multi-modal LLMs}, 
      author={Jinda Lu and Junkang Wu and Jinghan Li and Xiaojun Jia and Shuo Wang and YiFan Zhang and Junfeng Fang and Xiang Wang and Xiangnan He},
      year={2025},
      eprint={2502.01943},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2502.01943}, 
}

About

[ICML 2025] Official code of "DAMA: Data- and Model-aware Alignment of Multi-modal LLMs"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published