erikbeltran commited on
Commit
0e04af2
·
verified ·
1 Parent(s): e9c8f39

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +19 -11
app.py CHANGED
@@ -15,13 +15,13 @@ try:
15
  tokenizer = PreTrainedTokenizerFast.from_pretrained(MODEL_ID)
16
 
17
  # Ensure the tokenizer has the necessary special tokens
18
- #special_tokens = {
19
  # 'pad_token': '[PAD]',
20
- # 'eos_token': '</s>',
21
  # 'bos_token': '<s>',
22
  # 3 'unk_token': '<unk>'
23
- #}
24
- #tokenizer.add_special_tokens(special_tokens)
25
 
26
  model = AutoModelForCausalLM.from_pretrained(MODEL_ID, gguf_file=GGUF_FILE)
27
 
@@ -47,11 +47,19 @@ def format_diff_response(response):
47
  return '<br>'.join(formatted)
48
 
49
  def create_prompt(request, file_content, system_message):
50
- return f"""<system>{system_message}</system>
 
 
 
 
 
 
 
 
 
51
  <request>{request}</request>
52
- <file>
53
- {file_content}
54
- </file>"""
55
 
56
  @spaces.GPU
57
  def respond(request, file_content, system_message, max_tokens, temperature, top_p):
@@ -122,7 +130,7 @@ with gr.Blocks() as demo:
122
  max_tokens = gr.Slider(
123
  minimum=1,
124
  maximum=2048,
125
- value=512,
126
  step=1,
127
  label="Max Tokens"
128
  )
@@ -130,13 +138,13 @@ with gr.Blocks() as demo:
130
  minimum=0.1,
131
  maximum=4.0,
132
  value=0.7,
133
- step=0.1,
134
  label="Temperature"
135
  )
136
  top_p = gr.Slider(
137
  minimum=0.1,
138
  maximum=1.0,
139
- value=0.95,
140
  step=0.05,
141
  label="Top-p"
142
  )
 
15
  tokenizer = PreTrainedTokenizerFast.from_pretrained(MODEL_ID)
16
 
17
  # Ensure the tokenizer has the necessary special tokens
18
+ special_tokens = {
19
  # 'pad_token': '[PAD]',
20
+ 'eos_token': '<|eot_id|>'
21
  # 'bos_token': '<s>',
22
  # 3 'unk_token': '<unk>'
23
+ }
24
+ tokenizer.add_special_tokens(special_tokens)
25
 
26
  model = AutoModelForCausalLM.from_pretrained(MODEL_ID, gguf_file=GGUF_FILE)
27
 
 
47
  return '<br>'.join(formatted)
48
 
49
  def create_prompt(request, file_content, system_message):
50
+ # return f"""<system>{system_message}</system>
51
+ #<request>{request}</request>
52
+ #<file>{file_content}</file>"""
53
+ return f"""<|begin_of_text|><|start_header_id|>system<|end_header_id|>
54
+
55
+ Cutting Knowledge Date: December 2023
56
+ Today Date: 26 July 2024
57
+
58
+ {system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>
59
+
60
  <request>{request}</request>
61
+ #<file>{file_content}</file><|eot_id|><|start_header_id|>assistant<|end_header_id|>
62
+ """
 
63
 
64
  @spaces.GPU
65
  def respond(request, file_content, system_message, max_tokens, temperature, top_p):
 
130
  max_tokens = gr.Slider(
131
  minimum=1,
132
  maximum=2048,
133
+ value=128,
134
  step=1,
135
  label="Max Tokens"
136
  )
 
138
  minimum=0.1,
139
  maximum=4.0,
140
  value=0.7,
141
+ step=1,
142
  label="Temperature"
143
  )
144
  top_p = gr.Slider(
145
  minimum=0.1,
146
  maximum=1.0,
147
+ value=1,
148
  step=0.05,
149
  label="Top-p"
150
  )