Spaces:
Running
Running
updates
Browse files- app.py +1 -1
- src/about.py +13 -0
app.py
CHANGED
@@ -59,7 +59,7 @@ with gr.Blocks() as demo:
|
|
59 |
"""
|
60 |
<div style="text-align: center;">
|
61 |
<h1 style="font-size: 2.5rem; margin-bottom: 0.5rem;">TAG Leaderboard</h1>
|
62 |
-
<p style="font-size: 1.25rem; color: gray;">Evaluating complex natural language queries over
|
63 |
</div>
|
64 |
"""
|
65 |
)
|
|
|
59 |
"""
|
60 |
<div style="text-align: center;">
|
61 |
<h1 style="font-size: 2.5rem; margin-bottom: 0.5rem;">TAG Leaderboard</h1>
|
62 |
+
<p style="font-size: 1.25rem; color: gray;">Evaluating complex natural language queries over data.</p>
|
63 |
</div>
|
64 |
"""
|
65 |
)
|
src/about.py
CHANGED
@@ -35,6 +35,19 @@ In this leaderboard, you'll find execution accuracy comparisons of table questio
|
|
35 |
|
36 |
## How is accuracy measured?
|
37 |
Execution accuracy is measured as the number of exact matches to our annotated ground truth answers which are hand-labeled by experts.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
38 |
"""
|
39 |
|
40 |
EVALUATION_QUEUE_TEXT = """
|
|
|
35 |
|
36 |
## How is accuracy measured?
|
37 |
Execution accuracy is measured as the number of exact matches to our annotated ground truth answers which are hand-labeled by experts.
|
38 |
+
|
39 |
+
## Citation
|
40 |
+
```
|
41 |
+
@misc{{biswal2024text2sqlenoughunifyingai,
|
42 |
+
title={{Text2SQL is Not Enough: Unifying AI and Databases with TAG}},
|
43 |
+
author={{Asim Biswal and Liana Patel and Siddarth Jha and Amog Kamsetty and Shu Liu and Joseph E. Gonzalez and Carlos Guestrin and Matei Zaharia}},
|
44 |
+
year={2024},
|
45 |
+
eprint={2408.14717},
|
46 |
+
archivePrefix={{arXiv}},
|
47 |
+
primaryClass={{cs.DB}},
|
48 |
+
url={{https://arxiv.org/abs/2408.14717}},
|
49 |
+
}}
|
50 |
+
```
|
51 |
"""
|
52 |
|
53 |
EVALUATION_QUEUE_TEXT = """
|