When Gradient Optimization Is Not Enough: $dagger$ Dispersive and Anchoring Geometric Regularizer for Multimodal Learning
Multimodal learning aims to integrate complementary information from heterogeneous modalities, yet strong optimization alone does not guaranty well-structured representations. Even under carefully balanced training schemes, multimodal models often exhibit geometric pathologies, including intra-modal representation collapse and sample-level cross-modal inconsistency, which degrade both unimodal robustness and multimodal fusion. We identify representation geometry as a missing control axis in multimodal learning and propose regName, a lightweight geometry-aware regularization framework. regName enforces two complementary constraints on intermediate embeddings: an intra-modal dispersive […]