Safetensors
llava_llama

SymMPO: Mitigating Hallucination Through Theory-Consistent Symmetric Multimodal Preference Optimization

Wenqi Liu1, Xuemeng Song2, Jiaxi Li3, Yinwei Wei1, Zheng Na4, Jianhua Yin1, Liqiang Nie5
1Shandong University    2Southern University of Science and Technology    3University of Georgia   
4National University of Singapore    5Harbin Institute of Technology, Shenzhen   


Introduction

We present SymMPO, a framework for mitigating hallucination in multimodal large language models (MLLMs). Our method introduces a theory-consistent symmetric multimodal preference optimization approach that addresses the hallucination problem from a principled perspective. This repository provides the official implementation, pretrained checkpoints, and evaluation scripts built on top of LLaVA.

Citation

If you find our work helpful, please consider citing:

@inproceedings{
  liu2025mitigating,
  title={Mitigating Hallucination Through Theory-Consistent Symmetric Multimodal Preference Optimization},
  author={Wenqi Liu and Xuemeng Song and Jiaxi Li and Yinwei Wei and Na Zheng and Jianhua Yin and Liqiang Nie},
  booktitle={The Thirty-ninth Annual Conference on Neural Information Processing Systems},
  year={2025},
  url={https://openreview.net/forum?id=tIW29IpCwG}
}
Downloads last month
29
Safetensors
Model size
13B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for iLearn-Lab/NeurIPS25-SymMPO-13B

Finetuned
(5)
this model

Dataset used to train iLearn-Lab/NeurIPS25-SymMPO-13B

Paper for iLearn-Lab/NeurIPS25-SymMPO-13B