TogetherAI commited on
Commit
ce28b01
·
1 Parent(s): 7d4e07d

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +23 -27
app.py CHANGED
@@ -1,23 +1,24 @@
 
 
1
  from huggingface_hub import InferenceClient
2
  import gradio as gr
 
 
 
3
 
4
  client = InferenceClient(
5
- "HuggingFaceH4/zephyr-7b-alpha"
6
  )
7
 
8
  def format_prompt(message, history):
9
- prompt = "<s>"
10
- prompt += "[IDENDITY]You are Ailex, a clone and close collaborator of Einfach.Alex. As a part of the EinfachChat team, you assist your mentor Alex in a multitude of projects and initiatives. Your expertise is broad and encompasses sales, customer consulting, AI, Prompt Engineering, web design, and media design. Your life motto is 'Simply.Do!'. You communicate exclusively in German.[/IDENDITY]")
11
- for user_prompt, bot_response in history:
12
- prompt += f"[INST] {user_prompt} [/INST]"
13
- prompt += f" {bot_response}</s> "
14
- prompt += f"[INST] {message} [/INST]"
15
- return prompt
16
-
17
 
18
- def generate(
19
- prompt, history, temperature=0.9, max_new_tokens=512, top_p=0.95, repetition_penalty=1.0,
20
- ):
21
  temperature = float(temperature)
22
  if temperature < 1e-2:
23
  temperature = 1e-2
@@ -29,7 +30,7 @@ def generate(
29
  top_p=top_p,
30
  repetition_penalty=repetition_penalty,
31
  do_sample=True,
32
- seed=42,
33
  )
34
 
35
  formatted_prompt = format_prompt(prompt, history)
@@ -55,9 +56,9 @@ additional_inputs=[
55
  ),
56
  gr.Slider(
57
  label="Max new tokens",
58
- value=256,
59
- minimum=0,
60
- maximum=1048,
61
  step=64,
62
  interactive=True,
63
  info="The maximum numbers of new tokens",
@@ -82,22 +83,17 @@ additional_inputs=[
82
  )
83
  ]
84
 
85
- css = """
86
- #mkd {
87
- height: 500px;
88
- overflow: auto;
89
- border: 1px solid #ccc;
90
- }
91
  """
92
 
93
- with gr.Blocks(css=css, theme="NoCrypt/[email protected]") as demo: # Theme und CSS hier hinzugefügt
94
- gr.HTML("<h1><center>Chat with (Mistrailex 7B) <h1><center>")
95
- gr.HTML("<h3><center>Einfach.Fragen 💬<h3><center>")
96
- gr.HTML("<h3><center>Learn more about the model <a href='https://huggingface.co/docs/transformers/main/model_doc/mistral'>here</a>. 📚<h3><center>")
97
  gr.ChatInterface(
98
  generate,
99
  additional_inputs=additional_inputs,
100
- examples=[["What is the secret to life?"], ["Write me a recipe for pancakes."]]
101
  )
102
 
103
  demo.queue().launch(debug=True)
 
1
+
2
+
3
  from huggingface_hub import InferenceClient
4
  import gradio as gr
5
+ import random
6
+
7
+ API_URL = "https://api-inference.huggingface.co/models/"
8
 
9
  client = InferenceClient(
10
+ "mistralai/Mistral-7B-Instruct-v0.1"
11
  )
12
 
13
  def format_prompt(message, history):
14
+ prompt = "<s>You are Ailex, a clone and close collaborator of Einfach.Alex. As a part of the EinfachChat team, you assist your mentor Alex in a multitude of projects and initiatives. Your expertise is broad and encompasses sales, customer consulting, AI, Prompt Engineering, web design, and media design. Your life motto is 'Simply.Do!'. You communicate exclusively in German."
15
+ for user_prompt, bot_response in history:
16
+ prompt += f"[INST] {user_prompt} [/INST]"
17
+ prompt += f" {bot_response}</s> "
18
+ prompt += f"[INST] {message} [/INST]"
19
+ return prompt
 
 
20
 
21
+ def generate(prompt, history, temperature=0.9, max_new_tokens=512, top_p=0.95, repetition_penalty=1.0):
 
 
22
  temperature = float(temperature)
23
  if temperature < 1e-2:
24
  temperature = 1e-2
 
30
  top_p=top_p,
31
  repetition_penalty=repetition_penalty,
32
  do_sample=True,
33
+ seed=random.randint(0, 10**7),
34
  )
35
 
36
  formatted_prompt = format_prompt(prompt, history)
 
56
  ),
57
  gr.Slider(
58
  label="Max new tokens",
59
+ value=512,
60
+ minimum=64,
61
+ maximum=1024,
62
  step=64,
63
  interactive=True,
64
  info="The maximum numbers of new tokens",
 
83
  )
84
  ]
85
 
86
+ customCSS = """
87
+ #component-7 { # this is the default element ID of the chat component
88
+ height: 800px; # adjust the height as needed
89
+ flex-grow: 1;
90
+ }
 
91
  """
92
 
93
+ with gr.Blocks(css=NoCrypt/[email protected]) as demo:
 
 
 
94
  gr.ChatInterface(
95
  generate,
96
  additional_inputs=additional_inputs,
 
97
  )
98
 
99
  demo.queue().launch(debug=True)