Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -11,39 +11,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStream
|
|
11 |
MODEL_ID = "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
|
12 |
CHAT_TEMPLATE = "َAuto"
|
13 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
14 |
-
CONTEXT_LENGTH =
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
latex_delimiters_set = [{
|
19 |
-
"left": "\\(",
|
20 |
-
"right": "\\)",
|
21 |
-
"display": False
|
22 |
-
}, {
|
23 |
-
"left": "\\begin{equation}",
|
24 |
-
"right": "\\end{equation}",
|
25 |
-
"display": True
|
26 |
-
}, {
|
27 |
-
"left": "\\begin{align}",
|
28 |
-
"right": "\\end{align}",
|
29 |
-
"display": True
|
30 |
-
}, {
|
31 |
-
"left": "\\begin{alignat}",
|
32 |
-
"right": "\\end{alignat}",
|
33 |
-
"display": True
|
34 |
-
}, {
|
35 |
-
"left": "\\begin{gather}",
|
36 |
-
"right": "\\end{gather}",
|
37 |
-
"display": True
|
38 |
-
}, {
|
39 |
-
"left": "\\begin{CD}",
|
40 |
-
"right": "\\end{CD}",
|
41 |
-
"display": True
|
42 |
-
}, {
|
43 |
-
"left": "\\[",
|
44 |
-
"right": "\\]",
|
45 |
-
"display": True
|
46 |
-
}]
|
47 |
|
48 |
|
49 |
def predict(message, history, system_prompt, temperature, max_new_tokens, top_k, repetition_penalty, top_p):
|
@@ -100,7 +68,7 @@ gr.ChatInterface(
|
|
100 |
additional_inputs=[
|
101 |
gr.Textbox("You are a useful assistant. first recognize user request and then reply carfuly and thinking", label="System prompt"),
|
102 |
gr.Slider(0, 1, 0.6, label="Temperature"),
|
103 |
-
gr.Slider(0,
|
104 |
gr.Slider(1, 80, 40, label="Top K sampling"),
|
105 |
gr.Slider(0, 2, 1.1, label="Repetition penalty"),
|
106 |
gr.Slider(0, 1, 0.95, label="Top P sampling"),
|
|
|
11 |
MODEL_ID = "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
|
12 |
CHAT_TEMPLATE = "َAuto"
|
13 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
14 |
+
CONTEXT_LENGTH = 4096
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
|
17 |
def predict(message, history, system_prompt, temperature, max_new_tokens, top_k, repetition_penalty, top_p):
|
|
|
68 |
additional_inputs=[
|
69 |
gr.Textbox("You are a useful assistant. first recognize user request and then reply carfuly and thinking", label="System prompt"),
|
70 |
gr.Slider(0, 1, 0.6, label="Temperature"),
|
71 |
+
gr.Slider(0, 4096, 512, label="Max new tokens"),
|
72 |
gr.Slider(1, 80, 40, label="Top K sampling"),
|
73 |
gr.Slider(0, 2, 1.1, label="Repetition penalty"),
|
74 |
gr.Slider(0, 1, 0.95, label="Top P sampling"),
|