Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ configs:
|
|
23 |
## Dataset Description
|
24 |
|
25 |
<!-- Provide a longer summary of what this dataset is. -->
|
26 |
-
We introduce ViewSpatial-Bench
|
27 |
|
28 |
ViewSpatial-Bench addresses a critical gap: while VLMs excel at spatial reasoning from their own perspective, they struggle with perspective-taking—adopting another entity's spatial frame of reference—which is essential for embodied interaction and multi-agent collaboration.The figure below shows the construction pipeline and example demonstrations of our benchmark.
|
29 |
|
|
|
23 |
## Dataset Description
|
24 |
|
25 |
<!-- Provide a longer summary of what this dataset is. -->
|
26 |
+
We introduce **ViewSpatial-Bench**, a comprehensive benchmark with over 5,700 question-answer pairs across 1,000+ 3D scenes from ScanNet and MS-COCO validation sets. This benchmark evaluates VLMs' spatial localization capabilities from multiple perspectives, specifically testing both egocentric (camera) and allocentric (human subject) viewpoints across five distinct task types.
|
27 |
|
28 |
ViewSpatial-Bench addresses a critical gap: while VLMs excel at spatial reasoning from their own perspective, they struggle with perspective-taking—adopting another entity's spatial frame of reference—which is essential for embodied interaction and multi-agent collaboration.The figure below shows the construction pipeline and example demonstrations of our benchmark.
|
29 |
|