Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -24,7 +24,7 @@ def generate(
|
|
24 |
else:
|
25 |
prompt = f"<|im_start|>user\n{user_input}<|im_end|>\n<|im_start|>assistant\n"
|
26 |
outputs = pipe(prompt, max_new_tokens=max_new_tokens, do_sample=True,
|
27 |
-
temperature=temperature, top_k=top_k, top_p=top_p)
|
28 |
return outputs[0]["generated_text"]
|
29 |
|
30 |
model_choices = ["Locutusque/OpenCerebrum-1.0-7B-beta", "M4-ai/NeuralReyna-Mini-1.8B-v0.2", "Locutusque/Hyperion-3.0-Mistral-7B-DPO", "Locutusque/Hyperion-3.0-Mistral-7B-alpha", "M4-ai/tau-1.8B", "Locutusque/Hercules-4.0-Mistral-v0.2-7B", "Locutusque/Hercules-2.5-Mistral-7B", "M4-ai/tau-0.5B"]
|
|
|
24 |
else:
|
25 |
prompt = f"<|im_start|>user\n{user_input}<|im_end|>\n<|im_start|>assistant\n"
|
26 |
outputs = pipe(prompt, max_new_tokens=max_new_tokens, do_sample=True,
|
27 |
+
temperature=temperature, top_k=top_k, top_p=top_p, repetition_penalty=1.1)
|
28 |
return outputs[0]["generated_text"]
|
29 |
|
30 |
model_choices = ["Locutusque/OpenCerebrum-1.0-7B-beta", "M4-ai/NeuralReyna-Mini-1.8B-v0.2", "Locutusque/Hyperion-3.0-Mistral-7B-DPO", "Locutusque/Hyperion-3.0-Mistral-7B-alpha", "M4-ai/tau-1.8B", "Locutusque/Hercules-4.0-Mistral-v0.2-7B", "Locutusque/Hercules-2.5-Mistral-7B", "M4-ai/tau-0.5B"]
|