Jing Liu, Xinxin Zhu, Fei Liu, Longteng Guo, Zijia Zhao, Mingzhen Sun, Weining Wang, Hanqing Lu, Shiyu Zhou, Jiajun Zhang, Jinqiao Wang
In this paper, we propose an Omni-perception Pre-Trainer (OPT) for cross-modal understanding and generation, by jointly modeling visual, text and audio resources. OPT is constructed in an encoder-decoder framework, including three single-modal encoders to generate token-based embeddings for each modality, a cross-modal encoder to encode the correlations among the three modalities, and two cross-modal decoders to generate text and image respectively. For the OPT's pre-training, we design a multi-task pretext learning scheme to model multi-modal resources from three different data granularities, \ie, token-, modality-, and sample-level modeling, through which OPT learns to align and translate among different modalities. The pre-training task is carried out on a large amount of image-text-audio triplets from Open Images. Experimental results show that OPT can learn strong image-text-audio multi-modal representations and achieve promising results on a variety of cross-modal understanding and generation tasks.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Image Retrieval | Localized Narratives | Text-to-image R@1 | 0.4196 | OPT |
| Image Retrieval | Localized Narratives | Text-to-image R@10 | 0.8126 | OPT |
| Image Retrieval | Localized Narratives | Text-to-image R@5 | 0.72 | OPT |
| Text to Audio Retrieval | Localized Narratives | Text-to-audio R@1 | 0.78 | OPT |
| Text to Audio Retrieval | Localized Narratives | Text-to-audio R@10 | 0.958 | OPT |
| Text to Audio Retrieval | Localized Narratives | Text-to-audio R@5 | 0.927 | OPT |