meg-huggingface
commited on
Commit
·
3596f80
1
Parent(s):
95035be
Trying to make it work with new EAI versions
Browse files
src/backend/run_eval_suite_harness.py
CHANGED
@@ -46,6 +46,7 @@ def run_evaluation(eval_request: EvalRequest, task_names: list, num_fewshot: int
|
|
46 |
logger.info(f"Selected Tasks: {task_names}")
|
47 |
|
48 |
# no_cache=no_cache,
|
|
|
49 |
results = evaluator.simple_evaluate(
|
50 |
model="hf-causal-experimental", # "hf-causal"
|
51 |
model_args=eval_request.get_model_args(),
|
@@ -55,7 +56,6 @@ def run_evaluation(eval_request: EvalRequest, task_names: list, num_fewshot: int
|
|
55 |
device=device,
|
56 |
limit=limit,
|
57 |
write_out=True,
|
58 |
-
output_base_path="logs"
|
59 |
)
|
60 |
|
61 |
results["config"]["model_dtype"] = eval_request.precision
|
|
|
46 |
logger.info(f"Selected Tasks: {task_names}")
|
47 |
|
48 |
# no_cache=no_cache,
|
49 |
+
# output_base_path="logs"
|
50 |
results = evaluator.simple_evaluate(
|
51 |
model="hf-causal-experimental", # "hf-causal"
|
52 |
model_args=eval_request.get_model_args(),
|
|
|
56 |
device=device,
|
57 |
limit=limit,
|
58 |
write_out=True,
|
|
|
59 |
)
|
60 |
|
61 |
results["config"]["model_dtype"] = eval_request.precision
|