Abstract: Multimodal Large Language Models (MLLMs) excel in
various tasks, yet often struggle with modality bias, where
the model tends to rely heavily on a single modality and over-
look critical information in other modalities, which leads
to incorrect focus and generating irrelevant responses. In
this paper, we propose using the paradigm of preference
optimization to solve the modality bias problem, including
RLAIF-V-Bias, a debiased preference optimization dataset,
and a Noise-Aware Preference Optimization (NaPO) algo-
rithm. Specifically, we first construct the dataset by intro-
ducing perturbations to reduce the informational content of
certain modalities, compelling the model to rely on a specific
modality when generating negative responses. To address
the inevitable noise in automatically constructed data, we
combine the noise-robust Mean Absolute Error (MAE) with
the Binary Cross-Entropy (BCE) in Direct Preference Opti-
mization (DPO) by a negative Box-Cox transformation, and
dynamically adjust the algorithm’s noise robustness based
on the evaluated noise levels in the data. Extensive exper-
iments validate our approach, demonstrating not only its
effectiveness in mitigating modality bias but also its signifi-
cant role in minimizing hallucinations.
Loading