TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Visual Question Answering (VQA)/MMBench

Visual Question Answering (VQA) on MMBench

Metric: GPT-3.5 score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕GPT-3.5 score▼Extra DataPaperDate↕Code
1LLaVA-InternLM2-ViT + MoSLoRA73.8NoMixture-of-Subspaces in Low-Rank Adaptation2024-06-16Code
2CuMo-7B73NoCuMo: Scaling Multimodal LLM with Co-Upcycled Mi...2024-05-09Code
3LLaVA-LLaMA3-8B-ViT + MoSLoRA73NoMixture-of-Subspaces in Low-Rank Adaptation2024-06-16Code
4Video-LaVIT67.3NoVideo-LaVIT: Unified Video-Language Pre-training...2024-02-05Code
5DreamLLM-7B49.9NoDreamLLM: Synergistic Multimodal Comprehension a...2023-09-20Code