Choi Changin, Lim Sungjun, Rhee Wonjong
Retrieval-augmented generation can improve audio captioning by incorporating relevant audio-text pairs from a knowledge base. Existing methods typically rely solely on the input audio as a unimodal retrieval query. In contrast, we propose Generation-Assisted Multimodal Querying, which generates a text description of the input audio to enable multimodal querying. This approach aligns the query modality with the audio-text structure of the knowledge base, leading to more effective retrieval. Furthermore, we introduce a novel progressive learning strategy that gradually increases the number of interleaved audio-text pairs to enhance the training process. Our experiments on AudioCaps, Clotho, and Auto-ACD demonstrate that our approach achieves state-of-the-art results across these benchmarks.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Audio captioning | Clotho | BLEU-4 | 18.1 | MQ-Cap |
| Audio captioning | Clotho | CIDEr | 0.496 | MQ-Cap |
| Audio captioning | Clotho | METEOR | 0.192 | MQ-Cap |
| Audio captioning | Clotho | SPICE | 0.143 | MQ-Cap |
| Audio captioning | Clotho | SPIDEr | 0.319 | MQ-Cap |
| Audio captioning | AudioCaps | BLEU-4 | 0.301 | MQ-Cap |
| Audio captioning | AudioCaps | CIDEr | 0.845 | MQ-Cap |
| Audio captioning | AudioCaps | METEOR | 0.266 | MQ-Cap |
| Audio captioning | AudioCaps | SPICE | 0.194 | MQ-Cap |
| Audio captioning | AudioCaps | SPIDEr | 0.519 | MQ-Cap |