Spaces:
Running
Running
updates
Browse files- app.py +1 -1
- src/about.py +13 -0
app.py
CHANGED
|
@@ -59,7 +59,7 @@ with gr.Blocks() as demo:
|
|
| 59 |
"""
|
| 60 |
<div style="text-align: center;">
|
| 61 |
<h1 style="font-size: 2.5rem; margin-bottom: 0.5rem;">TAG Leaderboard</h1>
|
| 62 |
-
<p style="font-size: 1.25rem; color: gray;">Evaluating complex natural language queries over
|
| 63 |
</div>
|
| 64 |
"""
|
| 65 |
)
|
|
|
|
| 59 |
"""
|
| 60 |
<div style="text-align: center;">
|
| 61 |
<h1 style="font-size: 2.5rem; margin-bottom: 0.5rem;">TAG Leaderboard</h1>
|
| 62 |
+
<p style="font-size: 1.25rem; color: gray;">Evaluating complex natural language queries over data.</p>
|
| 63 |
</div>
|
| 64 |
"""
|
| 65 |
)
|
src/about.py
CHANGED
|
@@ -35,6 +35,19 @@ In this leaderboard, you'll find execution accuracy comparisons of table questio
|
|
| 35 |
|
| 36 |
## How is accuracy measured?
|
| 37 |
Execution accuracy is measured as the number of exact matches to our annotated ground truth answers which are hand-labeled by experts.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
"""
|
| 39 |
|
| 40 |
EVALUATION_QUEUE_TEXT = """
|
|
|
|
| 35 |
|
| 36 |
## How is accuracy measured?
|
| 37 |
Execution accuracy is measured as the number of exact matches to our annotated ground truth answers which are hand-labeled by experts.
|
| 38 |
+
|
| 39 |
+
## Citation
|
| 40 |
+
```
|
| 41 |
+
@misc{{biswal2024text2sqlenoughunifyingai,
|
| 42 |
+
title={{Text2SQL is Not Enough: Unifying AI and Databases with TAG}},
|
| 43 |
+
author={{Asim Biswal and Liana Patel and Siddarth Jha and Amog Kamsetty and Shu Liu and Joseph E. Gonzalez and Carlos Guestrin and Matei Zaharia}},
|
| 44 |
+
year={2024},
|
| 45 |
+
eprint={2408.14717},
|
| 46 |
+
archivePrefix={{arXiv}},
|
| 47 |
+
primaryClass={{cs.DB}},
|
| 48 |
+
url={{https://arxiv.org/abs/2408.14717}},
|
| 49 |
+
}}
|
| 50 |
+
```
|
| 51 |
"""
|
| 52 |
|
| 53 |
EVALUATION_QUEUE_TEXT = """
|