Spaces:
Running
Running
lixuejing
commited on
Commit
·
6d95172
1
Parent(s):
62da1e0
update
Browse files- app.py +15 -14
- src/about.py +36 -12
app.py
CHANGED
|
@@ -221,13 +221,13 @@ with demo:
|
|
| 221 |
elem_id="column-select",
|
| 222 |
interactive=True,
|
| 223 |
)
|
| 224 |
-
with gr.Row():
|
| 225 |
-
|
| 226 |
-
|
| 227 |
-
|
| 228 |
-
|
| 229 |
-
|
| 230 |
-
|
| 231 |
with gr.Column(min_width=320):
|
| 232 |
#with gr.Box(elem_id="box-filter"):
|
| 233 |
filter_columns_type = gr.CheckboxGroup(
|
|
@@ -349,13 +349,13 @@ with demo:
|
|
| 349 |
elem_id="column-select",
|
| 350 |
interactive=True,
|
| 351 |
)
|
| 352 |
-
with gr.Row():
|
| 353 |
-
|
| 354 |
-
|
| 355 |
-
|
| 356 |
-
|
| 357 |
-
|
| 358 |
-
|
| 359 |
with gr.Column(min_width=320):
|
| 360 |
#with gr.Box(elem_id="box-filter"):
|
| 361 |
filter_columns_type = gr.CheckboxGroup(
|
|
@@ -450,6 +450,7 @@ with demo:
|
|
| 450 |
leaderboard_table,
|
| 451 |
queue=True,
|
| 452 |
)
|
|
|
|
| 453 |
with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
|
| 454 |
gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
|
| 455 |
|
|
|
|
| 221 |
elem_id="column-select",
|
| 222 |
interactive=True,
|
| 223 |
)
|
| 224 |
+
#with gr.Row():
|
| 225 |
+
# hide_models = gr.CheckboxGroup(
|
| 226 |
+
# label="Hide models",
|
| 227 |
+
# choices = ["Private or deleted", "Contains a merge/moerge", "Flagged", "MoE"],
|
| 228 |
+
# value=[],
|
| 229 |
+
# interactive=True
|
| 230 |
+
# )
|
| 231 |
with gr.Column(min_width=320):
|
| 232 |
#with gr.Box(elem_id="box-filter"):
|
| 233 |
filter_columns_type = gr.CheckboxGroup(
|
|
|
|
| 349 |
elem_id="column-select",
|
| 350 |
interactive=True,
|
| 351 |
)
|
| 352 |
+
#with gr.Row():
|
| 353 |
+
# hide_models = gr.CheckboxGroup(
|
| 354 |
+
# label="Hide models",
|
| 355 |
+
# choices = ["Private or deleted", "Contains a merge/moerge", "Flagged", "MoE"],
|
| 356 |
+
# value=[],
|
| 357 |
+
# interactive=True
|
| 358 |
+
# )
|
| 359 |
with gr.Column(min_width=320):
|
| 360 |
#with gr.Box(elem_id="box-filter"):
|
| 361 |
filter_columns_type = gr.CheckboxGroup(
|
|
|
|
| 450 |
leaderboard_table,
|
| 451 |
queue=True,
|
| 452 |
)
|
| 453 |
+
gr.Markdown(EVALUATION_METRIC_TEXT, elem_classes="markdown-text")
|
| 454 |
with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
|
| 455 |
gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
|
| 456 |
|
src/about.py
CHANGED
|
@@ -25,23 +25,23 @@ class Tasks(Enum):
|
|
| 25 |
|
| 26 |
class Quotas(Enum):
|
| 27 |
Perception = Task("Perception", "overall", "Perception")
|
| 28 |
-
Perception_VisualGrounding = Task("Perception", "Visual Grounding", "
|
| 29 |
-
Perception_Counting = Task("Perception", "Counting", "
|
| 30 |
-
Perception_StateActivity_Understanding = Task("Perception", "State & Activity Understanding", "
|
| 31 |
SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
|
| 32 |
-
SpatialReasoning_Dynamic = Task("SpatialReasoning", "Dynamic", "
|
| 33 |
-
SpatialReasoning_Rd = Task("SpatialReasoning", "Relative direction", "
|
| 34 |
-
SpatialReasoning_mmm = Task("SpatialReasoning", "Multi-view matching", "
|
| 35 |
-
SpatialReasoning_red = Task("SpatialReasoning", "Relative distance", "
|
| 36 |
-
SpatialReasoning_ds = Task("SpatialReasoning", "Depth estimation", "
|
| 37 |
-
SpatialReasoning_rs = Task("SpatialReasoning", "Relative shape", "
|
| 38 |
-
SpatialReasoning_se = Task("SpatialReasoning", "Size estimation", "
|
| 39 |
Prediction = Task("Prediction", "overall", "Prediction")
|
| 40 |
Prediction_T = Task("Prediction", "Trajectory", "Prediction_Trajectory")
|
| 41 |
Prediction_F = Task("Prediction", "Future prediction", "Prediction_Future-prediction")
|
| 42 |
Planning = Task("Planning", "overall", "Planning")
|
| 43 |
-
Planning_G = Task("Planning", "Goal Decomposition", "
|
| 44 |
-
Planning_N = Task("Planning", "Navigation", "
|
| 45 |
|
| 46 |
NUM_FEWSHOT = 0 # Change with your few shot
|
| 47 |
# ---------------------------------------------------
|
|
@@ -109,6 +109,30 @@ You can find:
|
|
| 109 |
|
| 110 |
"""
|
| 111 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
EVALUATION_QUEUE_TEXT = """
|
| 113 |
## Evaluation Queue for the FlagEval VLM Leaderboard
|
| 114 |
Models added here will be automatically evaluated on the FlagEval cluster.
|
|
|
|
| 25 |
|
| 26 |
class Quotas(Enum):
|
| 27 |
Perception = Task("Perception", "overall", "Perception")
|
| 28 |
+
Perception_VisualGrounding = Task("Perception", "Visual Grounding", "P_VG")
|
| 29 |
+
Perception_Counting = Task("Perception", "Counting", "P_C")
|
| 30 |
+
Perception_StateActivity_Understanding = Task("Perception", "State & Activity Understanding", "P_SAU")
|
| 31 |
SpatialReasoning = Task("SpatialReasoning", "overall", "SpatialReasoning")
|
| 32 |
+
SpatialReasoning_Dynamic = Task("SpatialReasoning", "Dynamic", "SR_D")
|
| 33 |
+
SpatialReasoning_Rd = Task("SpatialReasoning", "Relative direction", "SR_R")
|
| 34 |
+
SpatialReasoning_mmm = Task("SpatialReasoning", "Multi-view matching", "SR_Mm")
|
| 35 |
+
SpatialReasoning_red = Task("SpatialReasoning", "Relative distance", "SR_Rd")
|
| 36 |
+
SpatialReasoning_ds = Task("SpatialReasoning", "Depth estimation", "SR_De")
|
| 37 |
+
SpatialReasoning_rs = Task("SpatialReasoning", "Relative shape", "SR_Rs")
|
| 38 |
+
SpatialReasoning_se = Task("SpatialReasoning", "Size estimation", "SR_Se")
|
| 39 |
Prediction = Task("Prediction", "overall", "Prediction")
|
| 40 |
Prediction_T = Task("Prediction", "Trajectory", "Prediction_Trajectory")
|
| 41 |
Prediction_F = Task("Prediction", "Future prediction", "Prediction_Future-prediction")
|
| 42 |
Planning = Task("Planning", "overall", "Planning")
|
| 43 |
+
Planning_G = Task("Planning", "Goal Decomposition", "P_GD")
|
| 44 |
+
Planning_N = Task("Planning", "Navigation", "P_N")
|
| 45 |
|
| 46 |
NUM_FEWSHOT = 0 # Change with your few shot
|
| 47 |
# ---------------------------------------------------
|
|
|
|
| 109 |
|
| 110 |
"""
|
| 111 |
|
| 112 |
+
|
| 113 |
+
EVALUATION_METRIC_TEXT = """
|
| 114 |
+
### 评测指标缩写介绍如下:
|
| 115 |
+
### Evaluation Metrics Abbreviations are introduced below:
|
| 116 |
+
Perception
|
| 117 |
+
- Perception_Visual Grounding(P_VG)
|
| 118 |
+
- Perception_Counting(P_C)
|
| 119 |
+
- Perception_State & Activity Understanding
|
| 120 |
+
SpatialReasoning
|
| 121 |
+
- SpatialReasoning_Dynamic(SR_D)
|
| 122 |
+
- SpatialReasoning_Relative direction(SR_Rd)
|
| 123 |
+
- SpatialReasoning_Multi-view matching(SR_Mm)
|
| 124 |
+
- SpatialReasoning_Relative distance(SR_Rd)
|
| 125 |
+
- SpatialReasoning_Depth estimation(SR_De)
|
| 126 |
+
- SpatialReasoning_Relative shape(SR_Rs)
|
| 127 |
+
- SpatialReasoning_Size estimation(SR_Se)
|
| 128 |
+
Prediction
|
| 129 |
+
- Prediction_Trajectory(P_T)
|
| 130 |
+
- Prediction_Futureprediction(P_Fd)
|
| 131 |
+
Planning
|
| 132 |
+
- Planning_Goal Decomposition(P_GD)
|
| 133 |
+
- Planning_Navigation(P_N)
|
| 134 |
+
"""
|
| 135 |
+
|
| 136 |
EVALUATION_QUEUE_TEXT = """
|
| 137 |
## Evaluation Queue for the FlagEval VLM Leaderboard
|
| 138 |
Models added here will be automatically evaluated on the FlagEval cluster.
|