lixuejing commited on
Commit
6d95172
·
1 Parent(s): 62da1e0
Files changed (2) hide show
  1. app.py +15 -14
  2. src/about.py +36 -12
app.py CHANGED
@@ -221,13 +221,13 @@ with demo:
221
  elem_id="column-select",
222
  interactive=True,
223
  )
224
- with gr.Row():
225
- hide_models = gr.CheckboxGroup(
226
- label="Hide models",
227
- choices = ["Private or deleted", "Contains a merge/moerge", "Flagged", "MoE"],
228
- value=[],
229
- interactive=True
230
- )
231
  with gr.Column(min_width=320):
232
  #with gr.Box(elem_id="box-filter"):
233
  filter_columns_type = gr.CheckboxGroup(
@@ -349,13 +349,13 @@ with demo:
349
  elem_id="column-select",
350
  interactive=True,
351
  )
352
- with gr.Row():
353
- hide_models = gr.CheckboxGroup(
354
- label="Hide models",
355
- choices = ["Private or deleted", "Contains a merge/moerge", "Flagged", "MoE"],
356
- value=[],
357
- interactive=True
358
- )
359
  with gr.Column(min_width=320):
360
  #with gr.Box(elem_id="box-filter"):
361
  filter_columns_type = gr.CheckboxGroup(
@@ -450,6 +450,7 @@ with demo:
450
  leaderboard_table,
451
  queue=True,
452
  )
 
453
  with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
454
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
455
 
 
221
  elem_id="column-select",
222
  interactive=True,
223
  )
224
+ #with gr.Row():
225
+ # hide_models = gr.CheckboxGroup(
226
+ # label="Hide models",
227
+ # choices = ["Private or deleted", "Contains a merge/moerge", "Flagged", "MoE"],
228
+ # value=[],
229
+ # interactive=True
230
+ # )
231
  with gr.Column(min_width=320):
232
  #with gr.Box(elem_id="box-filter"):
233
  filter_columns_type = gr.CheckboxGroup(
 
349
  elem_id="column-select",
350
  interactive=True,
351
  )
352
+ #with gr.Row():
353
+ # hide_models = gr.CheckboxGroup(
354
+ # label="Hide models",
355
+ # choices = ["Private or deleted", "Contains a merge/moerge", "Flagged", "MoE"],
356
+ # value=[],
357
+ # interactive=True
358
+ # )
359
  with gr.Column(min_width=320):
360
  #with gr.Box(elem_id="box-filter"):
361
  filter_columns_type = gr.CheckboxGroup(
 
450
  leaderboard_table,
451
  queue=True,
452
  )
453
+ gr.Markdown(EVALUATION_METRIC_TEXT, elem_classes="markdown-text")
454
  with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
455
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
456
 
src/about.py CHANGED
@@ -25,23 +25,23 @@ class Tasks(Enum):
25
 
26
  class Quotas(Enum):
27
  Perception = Task("Perception", "overall", "Perception")
28
- Perception_VisualGrounding = Task("Perception", "Visual Grounding", "Perception_VisualGrounding")
29
- Perception_Counting = Task("Perception", "Counting", "Perception_Counting")
30
- Perception_StateActivity_Understanding = Task("Perception", "State & Activity Understanding", "Perception_StateActivity-Understanding")
31
  SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
32
- SpatialReasoning_Dynamic = Task("SpatialReasoning", "Dynamic", "SpatialReasoning_Dynamic")
33
- SpatialReasoning_Rd = Task("SpatialReasoning", "Relative direction", "SpatialReasoning_Relative-direction")
34
- SpatialReasoning_mmm = Task("SpatialReasoning", "Multi-view matching", "SpatialReasoning_Multi-view-matching")
35
- SpatialReasoning_red = Task("SpatialReasoning", "Relative distance", "SpatialReasoning_Relative-distance")
36
- SpatialReasoning_ds = Task("SpatialReasoning", "Depth estimation", "SpatialReasoning_Depth-estimation")
37
- SpatialReasoning_rs = Task("SpatialReasoning", "Relative shape", "SpatialReasoning_Relative-shape")
38
- SpatialReasoning_se = Task("SpatialReasoning", "Size estimation", "SpatialReasoning_Size-estimation")
39
  Prediction = Task("Prediction", "overall", "Prediction")
40
  Prediction_T = Task("Prediction", "Trajectory", "Prediction_Trajectory")
41
  Prediction_F = Task("Prediction", "Future prediction", "Prediction_Future-prediction")
42
  Planning = Task("Planning", "overall", "Planning")
43
- Planning_G = Task("Planning", "Goal Decomposition", "Planning_Goal-Decomposition")
44
- Planning_N = Task("Planning", "Navigation", "Planning_Navigation")
45
 
46
  NUM_FEWSHOT = 0 # Change with your few shot
47
  # ---------------------------------------------------
@@ -109,6 +109,30 @@ You can find:
109
 
110
  """
111
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
112
  EVALUATION_QUEUE_TEXT = """
113
  ## Evaluation Queue for the FlagEval VLM Leaderboard
114
  Models added here will be automatically evaluated on the FlagEval cluster.
 
25
 
26
  class Quotas(Enum):
27
  Perception = Task("Perception", "overall", "Perception")
28
+ Perception_VisualGrounding = Task("Perception", "Visual Grounding", "P_VG")
29
+ Perception_Counting = Task("Perception", "Counting", "P_C")
30
+ Perception_StateActivity_Understanding = Task("Perception", "State & Activity Understanding", "P_SAU")
31
  SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
32
+ SpatialReasoning_Dynamic = Task("SpatialReasoning", "Dynamic", "SR_D")
33
+ SpatialReasoning_Rd = Task("SpatialReasoning", "Relative direction", "SR_R")
34
+ SpatialReasoning_mmm = Task("SpatialReasoning", "Multi-view matching", "SR_Mm")
35
+ SpatialReasoning_red = Task("SpatialReasoning", "Relative distance", "SR_Rd")
36
+ SpatialReasoning_ds = Task("SpatialReasoning", "Depth estimation", "SR_De")
37
+ SpatialReasoning_rs = Task("SpatialReasoning", "Relative shape", "SR_Rs")
38
+ SpatialReasoning_se = Task("SpatialReasoning", "Size estimation", "SR_Se")
39
  Prediction = Task("Prediction", "overall", "Prediction")
40
  Prediction_T = Task("Prediction", "Trajectory", "Prediction_Trajectory")
41
  Prediction_F = Task("Prediction", "Future prediction", "Prediction_Future-prediction")
42
  Planning = Task("Planning", "overall", "Planning")
43
+ Planning_G = Task("Planning", "Goal Decomposition", "P_GD")
44
+ Planning_N = Task("Planning", "Navigation", "P_N")
45
 
46
  NUM_FEWSHOT = 0 # Change with your few shot
47
  # ---------------------------------------------------
 
109
 
110
  """
111
 
112
+
113
+ EVALUATION_METRIC_TEXT = """
114
+ ### 评测指标缩写介绍如下:
115
+ ### Evaluation Metrics Abbreviations are introduced below:
116
+ Perception
117
+ - Perception_Visual Grounding(P_VG)
118
+ - Perception_Counting(P_C)
119
+ - Perception_State & Activity Understanding
120
+ SpatialReasoning
121
+ - SpatialReasoning_Dynamic(SR_D)
122
+ - SpatialReasoning_Relative direction(SR_Rd)
123
+ - SpatialReasoning_Multi-view matching(SR_Mm)
124
+ - SpatialReasoning_Relative distance(SR_Rd)
125
+ - SpatialReasoning_Depth estimation(SR_De)
126
+ - SpatialReasoning_Relative shape(SR_Rs)
127
+ - SpatialReasoning_Size estimation(SR_Se)
128
+ Prediction
129
+ - Prediction_Trajectory(P_T)
130
+ - Prediction_Futureprediction(P_Fd)
131
+ Planning
132
+ - Planning_Goal Decomposition(P_GD)
133
+ - Planning_Navigation(P_N)
134
+ """
135
+
136
  EVALUATION_QUEUE_TEXT = """
137
  ## Evaluation Queue for the FlagEval VLM Leaderboard
138
  Models added here will be automatically evaluated on the FlagEval cluster.