lixuejing commited on
Commit
8703807
·
1 Parent(s): a469435
Files changed (2) hide show
  1. src/about.py +1 -1
  2. src/leaderboard/read_evals.py +1 -1
src/about.py CHANGED
@@ -25,7 +25,7 @@ class Tasks(Enum):
25
 
26
  class Quotas(Enum):
27
  Perception = Task("Perception", "overall", "Perception")
28
- Perception_VisualGrounding = Task("Perception_VisualGrounding", "VisualGrounding", "VisualGrounding")
29
  SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
30
  Prediction = Task("Prediction", "overall", "Prediction")
31
  Planning = Task("Planning", "overall", "Planning")
 
25
 
26
  class Quotas(Enum):
27
  Perception = Task("Perception", "overall", "Perception")
28
+ Perception_VisualGrounding = Task("Perception", "VisualGrounding", "per_VisualGrounding")
29
  SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
30
  Prediction = Task("Prediction", "overall", "Prediction")
31
  Planning = Task("Planning", "overall", "Planning")
src/leaderboard/read_evals.py CHANGED
@@ -85,7 +85,7 @@ class EvalResult:
85
 
86
  for task in Quotas:
87
  task = task.value
88
-
89
  # We average all scores of a given metric (not all metrics are present in all files)
90
  accs = np.array([v.get(task.metric, None) for k, v in data["results"].items() if task.benchmark == k])
91
  if accs.size == 0 or any([acc is None for acc in accs]):
 
85
 
86
  for task in Quotas:
87
  task = task.value
88
+ print("task.value",task.value, "task.metric",task.metric, "task.benchmark", task.benchmark)
89
  # We average all scores of a given metric (not all metrics are present in all files)
90
  accs = np.array([v.get(task.metric, None) for k, v in data["results"].items() if task.benchmark == k])
91
  if accs.size == 0 or any([acc is None for acc in accs]):