Keywords: Multi-modal Fusion, Robustness, Inductive Priors, Synthetic Tasks, Small-Scale Experiments, Interpretability
TL;DR: we show that custom made fusion methods, that leverage domain knowledge, improve training time and accuracy with very limited data but harms robustness
Abstract: Using small-scale experiments with real and synthetic tasks, we compare multi-modal fusion methods, including a proposed `Product Fusion', to demonstrate how encoding task-specific priors affects performance. Our results highlight a crucial trade-off: aligning fusion design with priors boosts clean-data accuracy with limited data but significantly diminishes robustness to noisy inputs.
Code: zip
Submission Number: 29
Loading