AilexGPT commited on
Commit
1fdbc4a
·
1 Parent(s): e963c20

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +30 -30
app.py CHANGED
@@ -2,16 +2,17 @@ from huggingface_hub import InferenceClient
2
  import gradio as gr
3
 
4
  client = InferenceClient(
5
- "mistralai/Mixtral-8x7B-v0.1"
6
  )
7
 
 
8
  def format_prompt(message, history):
9
- prompt = "<s>"
10
- for user_prompt, bot_response in history:
11
- prompt += f"[INST] {user_prompt} [/INST]"
12
- prompt += f" {bot_response}</s> "
13
- prompt += f"[INST] {message} [/INST]"
14
- return prompt
15
 
16
  def generate(
17
  prompt, history, system_prompt, temperature=0.9, max_new_tokens=256, top_p=0.95, repetition_penalty=1.0,
@@ -39,65 +40,64 @@ def generate(
39
  yield output
40
  return output
41
 
 
42
  additional_inputs=[
43
  gr.Textbox(
44
- label="System-Prompt",
45
  max_lines=1,
46
  interactive=True,
47
  ),
48
  gr.Slider(
49
- label="Temperatur",
50
- value=0.5,
51
  minimum=0.0,
52
  maximum=1.0,
53
  step=0.05,
54
  interactive=True,
55
- info="Höhere Werte erzeugen vielfältigere Ergebnisse.",
56
  ),
57
  gr.Slider(
58
- label="Maximale Anzahl neuer Tokens",
59
- value=20480,
60
  minimum=0,
61
- maximum=32768,
62
  step=64,
63
  interactive=True,
64
- info="Maximale Anzahl an Tokens",
65
  ),
66
  gr.Slider(
67
- label="Top-p (Nucleus Sampling)",
68
- value=0.75,
69
  minimum=0.0,
70
  maximum=1,
71
  step=0.05,
72
  interactive=True,
73
- info="Höhere Werte ermöglichen eine größere Vielfalt",
74
  ),
75
  gr.Slider(
76
- label="Wiederholungsstrafe",
77
  value=1.2,
78
  minimum=1.0,
79
  maximum=2.0,
80
  step=0.05,
81
  interactive=True,
82
- info="Strafe für wiederholte Tokens",
83
  )
84
  ]
85
 
86
- examples=[
87
- ["", "Antworten Sie immer vollständig auf Englisch", 0.5, 20480, 0.75, 1.2],
88
- ["", "Répondez toujours complètement en Français", 0.5, 20480, 0.75, 1.2],
89
- ["", "Отвечай всегда полностью на русском языке", 0.5, 20480, 0.75, 1.2],
90
- ]
91
-
92
- description = r"""
93
- """
94
 
95
  gr.ChatInterface(
96
  fn=generate,
97
  chatbot=gr.Chatbot(show_label=False, show_share_button=False, show_copy_button=True, likeable=True, layout="panel"),
98
  additional_inputs=additional_inputs,
99
- title='Mixtral',
100
  examples=examples,
101
- description=description,
102
  concurrency_limit=20,
103
  ).launch(show_api=False)
 
2
  import gradio as gr
3
 
4
  client = InferenceClient(
5
+ "mistralai/Mixtral-8x7B-Instruct-v0.1"
6
  )
7
 
8
+
9
  def format_prompt(message, history):
10
+ prompt = "<s>"
11
+ for user_prompt, bot_response in history:
12
+ prompt += f"[INST] {user_prompt} [/INST]"
13
+ prompt += f" {bot_response}</s> "
14
+ prompt += f"[INST] {message} [/INST]"
15
+ return prompt
16
 
17
  def generate(
18
  prompt, history, system_prompt, temperature=0.9, max_new_tokens=256, top_p=0.95, repetition_penalty=1.0,
 
40
  yield output
41
  return output
42
 
43
+
44
  additional_inputs=[
45
  gr.Textbox(
46
+ label="System Prompt",
47
  max_lines=1,
48
  interactive=True,
49
  ),
50
  gr.Slider(
51
+ label="Temperature",
52
+ value=0.9,
53
  minimum=0.0,
54
  maximum=1.0,
55
  step=0.05,
56
  interactive=True,
57
+ info="Higher values produce more diverse outputs",
58
  ),
59
  gr.Slider(
60
+ label="Max new tokens",
61
+ value=256,
62
  minimum=0,
63
+ maximum=1048,
64
  step=64,
65
  interactive=True,
66
+ info="The maximum numbers of new tokens",
67
  ),
68
  gr.Slider(
69
+ label="Top-p (nucleus sampling)",
70
+ value=0.90,
71
  minimum=0.0,
72
  maximum=1,
73
  step=0.05,
74
  interactive=True,
75
+ info="Higher values sample more low-probability tokens",
76
  ),
77
  gr.Slider(
78
+ label="Repetition penalty",
79
  value=1.2,
80
  minimum=1.0,
81
  maximum=2.0,
82
  step=0.05,
83
  interactive=True,
84
+ info="Penalize repeated tokens",
85
  )
86
  ]
87
 
88
+ examples=[["I'm planning a vacation to Japan. Can you suggest a one-week itinerary including must-visit places and local cuisines to try?", None, None, None, None, None, ],
89
+ ["Can you write a short story about a time-traveling detective who solves historical mysteries?", None, None, None, None, None,],
90
+ ["I'm trying to learn French. Can you provide some common phrases that would be useful for a beginner, along with their pronunciations?", None, None, None, None, None,],
91
+ ["I have chicken, rice, and bell peppers in my kitchen. Can you suggest an easy recipe I can make with these ingredients?", None, None, None, None, None,],
92
+ ["Can you explain how the QuickSort algorithm works and provide a Python implementation?", None, None, None, None, None,],
93
+ ["What are some unique features of Rust that make it stand out compared to other systems programming languages like C++?", None, None, None, None, None,],
94
+ ]
 
95
 
96
  gr.ChatInterface(
97
  fn=generate,
98
  chatbot=gr.Chatbot(show_label=False, show_share_button=False, show_copy_button=True, likeable=True, layout="panel"),
99
  additional_inputs=additional_inputs,
100
+ title="Mixtral 46.7B",
101
  examples=examples,
 
102
  concurrency_limit=20,
103
  ).launch(show_api=False)