Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Clémentine
commited on
Commit
·
23f614e
1
Parent(s):
459932d
simplified env vars
Browse files- README.md +0 -1
- app.py +4 -3
- src/envs.py +2 -2
- src/submission/submit.py +2 -2
README.md
CHANGED
|
@@ -15,7 +15,6 @@ space_ci:
|
|
| 15 |
private: true
|
| 16 |
secrets:
|
| 17 |
- HF_TOKEN
|
| 18 |
-
- H4_TOKEN
|
| 19 |
tags:
|
| 20 |
- leaderboard
|
| 21 |
short_description: Track, rank and evaluate open LLMs and chatbots
|
|
|
|
| 15 |
private: true
|
| 16 |
secrets:
|
| 17 |
- HF_TOKEN
|
|
|
|
| 18 |
tags:
|
| 19 |
- leaderboard
|
| 20 |
short_description: Track, rank and evaluate open LLMs and chatbots
|
app.py
CHANGED
|
@@ -32,9 +32,10 @@ from src.envs import (
|
|
| 32 |
API,
|
| 33 |
EVAL_REQUESTS_PATH,
|
| 34 |
AGGREGATED_REPO,
|
| 35 |
-
|
| 36 |
QUEUE_REPO,
|
| 37 |
REPO_ID,
|
|
|
|
| 38 |
)
|
| 39 |
from src.populate import get_evaluation_queue_df, get_leaderboard_df
|
| 40 |
from src.submission.submit import add_new_eval
|
|
@@ -48,7 +49,7 @@ enable_space_ci()
|
|
| 48 |
|
| 49 |
|
| 50 |
def restart_space():
|
| 51 |
-
API.restart_space(repo_id=REPO_ID, token=
|
| 52 |
|
| 53 |
|
| 54 |
def time_diff_wrapper(func):
|
|
@@ -98,7 +99,7 @@ def init_space(full_init: bool = True):
|
|
| 98 |
restart_space()
|
| 99 |
|
| 100 |
# Always retrieve the leaderboard DataFrame
|
| 101 |
-
leaderboard_dataset = datasets.load_dataset(AGGREGATED_REPO, "default", split="train")
|
| 102 |
leaderboard_df = get_leaderboard_df(
|
| 103 |
leaderboard_dataset=leaderboard_dataset,
|
| 104 |
cols=COLS,
|
|
|
|
| 32 |
API,
|
| 33 |
EVAL_REQUESTS_PATH,
|
| 34 |
AGGREGATED_REPO,
|
| 35 |
+
HF_TOKEN,
|
| 36 |
QUEUE_REPO,
|
| 37 |
REPO_ID,
|
| 38 |
+
HF_HOME,
|
| 39 |
)
|
| 40 |
from src.populate import get_evaluation_queue_df, get_leaderboard_df
|
| 41 |
from src.submission.submit import add_new_eval
|
|
|
|
| 49 |
|
| 50 |
|
| 51 |
def restart_space():
|
| 52 |
+
API.restart_space(repo_id=REPO_ID, token=HF_TOKEN)
|
| 53 |
|
| 54 |
|
| 55 |
def time_diff_wrapper(func):
|
|
|
|
| 99 |
restart_space()
|
| 100 |
|
| 101 |
# Always retrieve the leaderboard DataFrame
|
| 102 |
+
leaderboard_dataset = datasets.load_dataset(AGGREGATED_REPO, "default", split="train", cache_dir=HF_HOME)
|
| 103 |
leaderboard_df = get_leaderboard_df(
|
| 104 |
leaderboard_dataset=leaderboard_dataset,
|
| 105 |
cols=COLS,
|
src/envs.py
CHANGED
|
@@ -2,7 +2,7 @@ import os
|
|
| 2 |
from huggingface_hub import HfApi
|
| 3 |
|
| 4 |
# clone / pull the lmeh eval data
|
| 5 |
-
|
| 6 |
|
| 7 |
REPO_ID = "HuggingFaceH4/open_llm_leaderboard"
|
| 8 |
QUEUE_REPO = "open-llm-leaderboard/requests"
|
|
@@ -29,4 +29,4 @@ RATE_LIMIT_PERIOD = 7
|
|
| 29 |
RATE_LIMIT_QUOTA = 5
|
| 30 |
HAS_HIGHER_RATE_LIMIT = ["TheBloke"]
|
| 31 |
|
| 32 |
-
API = HfApi(token=
|
|
|
|
| 2 |
from huggingface_hub import HfApi
|
| 3 |
|
| 4 |
# clone / pull the lmeh eval data
|
| 5 |
+
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
| 6 |
|
| 7 |
REPO_ID = "HuggingFaceH4/open_llm_leaderboard"
|
| 8 |
QUEUE_REPO = "open-llm-leaderboard/requests"
|
|
|
|
| 29 |
RATE_LIMIT_QUOTA = 5
|
| 30 |
HAS_HIGHER_RATE_LIMIT = ["TheBloke"]
|
| 31 |
|
| 32 |
+
API = HfApi(token=HF_TOKEN)
|
src/submission/submit.py
CHANGED
|
@@ -8,7 +8,7 @@ from src.display.formatting import styled_error, styled_message, styled_warning
|
|
| 8 |
from src.envs import (
|
| 9 |
API,
|
| 10 |
EVAL_REQUESTS_PATH,
|
| 11 |
-
|
| 12 |
QUEUE_REPO,
|
| 13 |
RATE_LIMIT_PERIOD,
|
| 14 |
RATE_LIMIT_QUOTA,
|
|
@@ -76,7 +76,7 @@ def add_new_eval(
|
|
| 76 |
# Is the model on the hub?
|
| 77 |
if weight_type in ["Delta", "Adapter"]:
|
| 78 |
base_model_on_hub, error, _ = is_model_on_hub(
|
| 79 |
-
model_name=base_model, revision=revision, token=
|
| 80 |
)
|
| 81 |
if not base_model_on_hub:
|
| 82 |
return styled_error(f'Base model "{base_model}" {error}')
|
|
|
|
| 8 |
from src.envs import (
|
| 9 |
API,
|
| 10 |
EVAL_REQUESTS_PATH,
|
| 11 |
+
HF_TOKEN,
|
| 12 |
QUEUE_REPO,
|
| 13 |
RATE_LIMIT_PERIOD,
|
| 14 |
RATE_LIMIT_QUOTA,
|
|
|
|
| 76 |
# Is the model on the hub?
|
| 77 |
if weight_type in ["Delta", "Adapter"]:
|
| 78 |
base_model_on_hub, error, _ = is_model_on_hub(
|
| 79 |
+
model_name=base_model, revision=revision, token=HF_TOKEN, test_tokenizer=True
|
| 80 |
)
|
| 81 |
if not base_model_on_hub:
|
| 82 |
return styled_error(f'Base model "{base_model}" {error}')
|