Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -15,13 +15,13 @@ try:
|
|
15 |
tokenizer = PreTrainedTokenizerFast.from_pretrained(MODEL_ID)
|
16 |
|
17 |
# Ensure the tokenizer has the necessary special tokens
|
18 |
-
|
19 |
# 'pad_token': '[PAD]',
|
20 |
-
|
21 |
# 'bos_token': '<s>',
|
22 |
# 3 'unk_token': '<unk>'
|
23 |
-
|
24 |
-
|
25 |
|
26 |
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, gguf_file=GGUF_FILE)
|
27 |
|
@@ -47,11 +47,19 @@ def format_diff_response(response):
|
|
47 |
return '<br>'.join(formatted)
|
48 |
|
49 |
def create_prompt(request, file_content, system_message):
|
50 |
-
return f"""<system>{system_message}</system>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
<request>{request}</request>
|
52 |
-
|
53 |
-
|
54 |
-
</file>"""
|
55 |
|
56 |
@spaces.GPU
|
57 |
def respond(request, file_content, system_message, max_tokens, temperature, top_p):
|
@@ -122,7 +130,7 @@ with gr.Blocks() as demo:
|
|
122 |
max_tokens = gr.Slider(
|
123 |
minimum=1,
|
124 |
maximum=2048,
|
125 |
-
value=
|
126 |
step=1,
|
127 |
label="Max Tokens"
|
128 |
)
|
@@ -130,13 +138,13 @@ with gr.Blocks() as demo:
|
|
130 |
minimum=0.1,
|
131 |
maximum=4.0,
|
132 |
value=0.7,
|
133 |
-
step=
|
134 |
label="Temperature"
|
135 |
)
|
136 |
top_p = gr.Slider(
|
137 |
minimum=0.1,
|
138 |
maximum=1.0,
|
139 |
-
value=
|
140 |
step=0.05,
|
141 |
label="Top-p"
|
142 |
)
|
|
|
15 |
tokenizer = PreTrainedTokenizerFast.from_pretrained(MODEL_ID)
|
16 |
|
17 |
# Ensure the tokenizer has the necessary special tokens
|
18 |
+
special_tokens = {
|
19 |
# 'pad_token': '[PAD]',
|
20 |
+
'eos_token': '<|eot_id|>'
|
21 |
# 'bos_token': '<s>',
|
22 |
# 3 'unk_token': '<unk>'
|
23 |
+
}
|
24 |
+
tokenizer.add_special_tokens(special_tokens)
|
25 |
|
26 |
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, gguf_file=GGUF_FILE)
|
27 |
|
|
|
47 |
return '<br>'.join(formatted)
|
48 |
|
49 |
def create_prompt(request, file_content, system_message):
|
50 |
+
# return f"""<system>{system_message}</system>
|
51 |
+
#<request>{request}</request>
|
52 |
+
#<file>{file_content}</file>"""
|
53 |
+
return f"""<|begin_of_text|><|start_header_id|>system<|end_header_id|>
|
54 |
+
|
55 |
+
Cutting Knowledge Date: December 2023
|
56 |
+
Today Date: 26 July 2024
|
57 |
+
|
58 |
+
{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>
|
59 |
+
|
60 |
<request>{request}</request>
|
61 |
+
#<file>{file_content}</file><|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
62 |
+
"""
|
|
|
63 |
|
64 |
@spaces.GPU
|
65 |
def respond(request, file_content, system_message, max_tokens, temperature, top_p):
|
|
|
130 |
max_tokens = gr.Slider(
|
131 |
minimum=1,
|
132 |
maximum=2048,
|
133 |
+
value=128,
|
134 |
step=1,
|
135 |
label="Max Tokens"
|
136 |
)
|
|
|
138 |
minimum=0.1,
|
139 |
maximum=4.0,
|
140 |
value=0.7,
|
141 |
+
step=1,
|
142 |
label="Temperature"
|
143 |
)
|
144 |
top_p = gr.Slider(
|
145 |
minimum=0.1,
|
146 |
maximum=1.0,
|
147 |
+
value=1,
|
148 |
step=0.05,
|
149 |
label="Top-p"
|
150 |
)
|