Spaces:
Running
Running
lixuejing
commited on
Commit
·
8703807
1
Parent(s):
a469435
update
Browse files- src/about.py +1 -1
- src/leaderboard/read_evals.py +1 -1
src/about.py
CHANGED
@@ -25,7 +25,7 @@ class Tasks(Enum):
|
|
25 |
|
26 |
class Quotas(Enum):
|
27 |
Perception = Task("Perception", "overall", "Perception")
|
28 |
-
Perception_VisualGrounding = Task("
|
29 |
SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
|
30 |
Prediction = Task("Prediction", "overall", "Prediction")
|
31 |
Planning = Task("Planning", "overall", "Planning")
|
|
|
25 |
|
26 |
class Quotas(Enum):
|
27 |
Perception = Task("Perception", "overall", "Perception")
|
28 |
+
Perception_VisualGrounding = Task("Perception", "VisualGrounding", "per_VisualGrounding")
|
29 |
SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
|
30 |
Prediction = Task("Prediction", "overall", "Prediction")
|
31 |
Planning = Task("Planning", "overall", "Planning")
|
src/leaderboard/read_evals.py
CHANGED
@@ -85,7 +85,7 @@ class EvalResult:
|
|
85 |
|
86 |
for task in Quotas:
|
87 |
task = task.value
|
88 |
-
|
89 |
# We average all scores of a given metric (not all metrics are present in all files)
|
90 |
accs = np.array([v.get(task.metric, None) for k, v in data["results"].items() if task.benchmark == k])
|
91 |
if accs.size == 0 or any([acc is None for acc in accs]):
|
|
|
85 |
|
86 |
for task in Quotas:
|
87 |
task = task.value
|
88 |
+
print("task.value",task.value, "task.metric",task.metric, "task.benchmark", task.benchmark)
|
89 |
# We average all scores of a given metric (not all metrics are present in all files)
|
90 |
accs = np.array([v.get(task.metric, None) for k, v in data["results"].items() if task.benchmark == k])
|
91 |
if accs.size == 0 or any([acc is None for acc in accs]):
|