schuler commited on
Commit
96a08ea
·
verified ·
1 Parent(s): cc932be

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +42 -30
app.py CHANGED
@@ -17,6 +17,11 @@ def load_model(repo_name):
17
 
18
  tokenizer, generator_conf, model = load_model(REPO_NAME)
19
 
 
 
 
 
 
20
 
21
  def respond(
22
  message,
@@ -26,37 +31,44 @@ def respond(
26
  temperature,
27
  top_p,
28
  ):
29
- messages = [{"role": "system", "content": system_message}]
30
-
31
- for val in history:
32
- if val[0]:
33
- messages.append({"role": "user", "content": val[0]})
34
- if val[1]:
35
- messages.append({"role": "assistant", "content": val[1]})
36
-
37
- messages.append({"role": "user", "content": message})
38
-
39
- for message in messages:
40
- role = "<|assistant|>" if message['role'] == 'assistant' else "<|user|>"
41
- prompt += f"\n{role}\n{message['content']}\n<|end|>\n"
42
- # prompt += f"\n<|user|>\n{user_text}\n<|end|><|assistant|>\n"
43
-
44
- # Generate the response
45
- response_output = generator(
46
- prompt,
47
- generation_config=generator_conf,
48
- max_new_tokens=64,
49
- do_sample=True,
50
- top_p=0.25,
51
- repetition_penalty=1.2
52
- )
53
-
54
- generated_text = response_output[0]['generated_text']
 
 
 
 
 
 
 
 
 
55
 
56
- # st.session_state.last_response = generated_text
57
 
58
- # Extract the assistant's response
59
- yield generated_text[len(prompt):].strip()
60
 
61
  """
62
  for message in client.chat_completion(
@@ -79,7 +91,7 @@ For information on how to customize the ChatInterface, peruse the gradio docs: h
79
  demo = gr.ChatInterface(
80
  respond,
81
  additional_inputs=[
82
- gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
83
  gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
84
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
85
  gr.Slider(
 
17
 
18
  tokenizer, generator_conf, model = load_model(REPO_NAME)
19
 
20
+ global_error = ''
21
+ try:
22
+ generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
23
+ except Exception as e:
24
+ global_error = f"Failed to load model: {str(e)}"
25
 
26
  def respond(
27
  message,
 
31
  temperature,
32
  top_p,
33
  ):
34
+ result = 'none'
35
+ try:
36
+ messages = [{"role": "system", "content": system_message}]
37
+
38
+ for val in history:
39
+ if val[0]:
40
+ messages.append({"role": "user", "content": val[0]})
41
+ if val[1]:
42
+ messages.append({"role": "assistant", "content": val[1]})
43
+
44
+ messages.append({"role": "user", "content": message})
45
+
46
+ for message in messages:
47
+ role = "<|assistant|>" if message['role'] == 'assistant' else "<|user|>"
48
+ prompt += f"\n{role}\n{message['content']}\n<|end|>\n"
49
+ # prompt += f"\n<|user|>\n{user_text}\n<|end|><|assistant|>\n"
50
+
51
+ # Generate the response
52
+ response_output = generator(
53
+ prompt,
54
+ generation_config=generator_conf,
55
+ max_new_tokens=64,
56
+ do_sample=True,
57
+ top_p=0.25,
58
+ repetition_penalty=1.2
59
+ )
60
+
61
+ generated_text = response_output[0]['generated_text']
62
+
63
+ # st.session_state.last_response = generated_text
64
+
65
+ # Extract the assistant's response
66
+ result = generated_text[len(prompt):].strip()
67
+ except Exception as error:
68
+ result = str(error)
69
 
70
+ yield result
71
 
 
 
72
 
73
  """
74
  for message in client.chat_completion(
 
91
  demo = gr.ChatInterface(
92
  respond,
93
  additional_inputs=[
94
+ gr.Textbox(value="You are a friendly Chatbot." + global_error, label="System message"),
95
  gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
96
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
97
  gr.Slider(