Yuncheng Guo, Xiaodong Gu
Large-scale pre-trained Vision-Language Models (VLMs) have significantly advanced transfer learning across diverse tasks. However, adapting these models with limited few-shot data often leads to overfitting, undermining their ability to generalize to new tasks. To address this, we propose Multi-Modal Representation Learning (MMRL), which introduces a shared, learnable, modality-agnostic representation space. MMRL generates space tokens projected into both text and image encoders as representation tokens, enabling more effective cross-modal interactions. Unlike prior methods that mainly optimize class token features, MMRL inserts representation tokens into higher encoder layers--where task-specific features are more prominent--while preserving general knowledge in the lower layers. During training, both class and representation features are jointly optimized: a trainable projection layer is applied to representation tokens for task adaptation, while the projection layer for class token remains frozen to retain pre-trained knowledge. To further promote generalization, we introduce a regularization term aligning class and text features with the frozen VLM's zero-shot features. At inference, a decoupling strategy uses both class and representation features for base tasks, but only class features for novel tasks due to their stronger generalization. Building upon this, we propose MMRL++, a parameter-efficient and interaction-aware extension that significantly reduces trainable parameters and enhances intra-modal interactions--particularly across the layers of representation tokens--allowing gradient sharing and instance-specific information to propagate more effectively through the network. Extensive experiments on 15 datasets demonstrate that MMRL and MMRL++ consistently outperform state-of-the-art methods, achieving a strong balance between task-specific adaptation and generalization.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Prompt Engineering | Stanford Cars | Harmonic mean | 78.18 | MMRL++ |
| Prompt Engineering | Oxford 102 Flower | Harmonic mean | 87.01 | MMRL++ |
| Prompt Engineering | EuroSAT | Harmonic mean | 91.94 | MMRL++ |
| Prompt Engineering | Oxford-IIIT Pet Dataset | Harmonic mean | 96.51 | MMRL++ |
| Prompt Engineering | DTD | Harmonic mean | 74.46 | MMRL++ |
| Prompt Engineering | UCF101 | Harmonic mean | 83.81 | MMRL++ |
| Prompt Engineering | Food-101 | Harmonic mean | 91.1 | MMRL++ |
| Prompt Engineering | Caltech-101 | Harmonic mean | 96.75 | MMRL++ |
| Prompt Engineering | ImageNet | Harmonic mean | 74.44 | MMRL++ |
| Prompt Engineering | FGVC-Aircraft | Harmonic mean | 42.24 | MMRL++ |
| Prompt Engineering | SUN397 | Harmonic mean | 81.28 | MMRL++ |