TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Computer Vision/Visual Grounding/RefCOCO+ testA

Visual Grounding on RefCOCO+ testA

Metric: Accuracy (%) (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Accuracy (%)▼Extra DataPaperDate↕Code
1Florence-2-large-ft95.3YesFlorence-2: Advancing a Unified Representation f...2023-11-10Code
2mPLUG-292.8NomPLUG-2: A Modularized Multi-modal Foundation Mo...2023-02-01Code
3X2-VLM (large)92.1NoX$^2$-VLM: All-In-One Pre-trained Model For Visi...2022-11-22Code
4XFM (base)90.4NoToward Building General Foundation Models for La...2023-01-12Code
5X2-VLM (base)90.3NoX$^2$-VLM: All-In-One Pre-trained Model For Visi...2022-11-22Code
6X-VLM (base)89NoMulti-Grained Vision Language Pre-Training: Alig...2021-11-16Code