TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Visual Question Answering (VQA)/VQA v2 test-std

Visual Question Answering (VQA) on VQA v2 test-std

Metric: other (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕other▼Extra DataPaperDate↕Code
1mPLUG-Huge77.02NomPLUG: Effective and Efficient Vision-Language L...2022-05-24Code
2ONE-PEACE74.15NoONE-PEACE: Exploring One General Representation ...2023-05-18Code
3OFA73.35NoOFA: Unifying Architectures, Tasks, and Modaliti...2022-02-07Code
4VLMo72.87NoVLMo: Unified Vision-Language Pre-Training with ...2021-11-03Code
5Prismer69.7NoPrismer: A Vision-Language Model with Multi-Task...2023-03-04Code
6MSR + MS Cog. Svcs., X10 models67.87NoVinVL: Revisiting Visual Representations in Visi...2021-01-02Code
7MSR + MS Cog. Svcs.66.68NoVinVL: Revisiting Visual Representations in Visi...2021-01-02Code
8BGN, ensemble66.28NoBilinear Graph Networks for Visual Question Answ...2019-07-23-
9ERNIE-ViL-single model65.24NoERNIE-ViL: Knowledge Enhanced Vision-Language Re...2020-06-30-
10Single, w/o VLP64.77NoIn Defense of Grid Features for Visual Question ...2020-01-10Code
11Single, w/o VLP63.78NoDeep Multimodal Neural Architecture Search2020-04-25Code