TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Visual Question Answering (VQA)/6-DoF SpatialBench

Visual Question Answering (VQA) on 6-DoF SpatialBench

Metric: Total (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Total▼Extra DataPaperDate↕Code
1SoFar43.9NoSoFar: Language-Grounded Orientation Bridges Spa...2025-02-18Code
2GPT-4o36.2NoGPT-4o System Card2024-10-25-
3RoboPoint33.5NoRoboPoint: A Vision-Language Model for Spatial A...2024-06-15-
4SpatialBot32.7NoSpatialBot: Precise Spatial Understanding with V...2024-06-19Code
5SpaceMantis28.9NoSpatialVLM: Endowing Vision-Language Models with...2024-01-22-
6SpaceLLaVA28.2NoSpatialVLM: Endowing Vision-Language Models with...2024-01-22-
7LLaVA-1.527.2NoImproved Baselines with Visual Instruction Tuning2023-10-05Code