TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Visual Question Answering (VQA)/V*bench

Visual Question Answering (VQA) on V*bench

Metric: Accuracy (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Accuracy▼Extra DataPaperDate↕Code
1LLaVA-OneVision7B w. FOCUS92.15No---
2LLaVA-OneVision7B w. ZoomEye90.58NoZoomEye: Enhancing Multimodal LLMs with Human-Li...2024-11-25Code
3IVM-Enhanced GPT4-V81.2NoInstruction-Guided Visual Masking2024-05-30Code
4SEAL75.39YesV*: Guided Visual Search as a Core Mechanism in ...2023-12-21Code
5LLaVA-OneVision7B74.46NoLLaVA-OneVision: Easy Visual Task Transfer2024-08-06Code