TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Visual Question Answering (VQA)/6-DoF SpatialBench

Visual Question Answering (VQA) on 6-DoF SpatialBench

Metric: Position-rel (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Position-rel▼Extra DataPaperDate↕Code
1SoFar59.6NoSoFar: Language-Grounded Orientation Bridges Spa...2025-02-18Code
2SpatialBot50.9NoSpatialBot: Precise Spatial Understanding with V...2024-06-19Code
3GPT-4o49.4NoGPT-4o System Card2024-10-25-
4RoboPoint43.8NoRoboPoint: A Vision-Language Model for Spatial A...2024-06-15-
5SpaceMantis33.6NoSpatialVLM: Endowing Vision-Language Models with...2024-01-22-
6SpaceLLaVA32.4NoSpatialVLM: Endowing Vision-Language Models with...2024-01-22-
7LLaVA-1.530.9NoImproved Baselines with Visual Instruction Tuning2023-10-05Code