Spaces:
Sleeping
Sleeping
Commit
·
ce28b01
1
Parent(s):
7d4e07d
Update app.py
Browse files
app.py
CHANGED
@@ -1,23 +1,24 @@
|
|
|
|
|
|
1 |
from huggingface_hub import InferenceClient
|
2 |
import gradio as gr
|
|
|
|
|
|
|
3 |
|
4 |
client = InferenceClient(
|
5 |
-
"
|
6 |
)
|
7 |
|
8 |
def format_prompt(message, history):
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
return prompt
|
16 |
-
|
17 |
|
18 |
-
def generate(
|
19 |
-
prompt, history, temperature=0.9, max_new_tokens=512, top_p=0.95, repetition_penalty=1.0,
|
20 |
-
):
|
21 |
temperature = float(temperature)
|
22 |
if temperature < 1e-2:
|
23 |
temperature = 1e-2
|
@@ -29,7 +30,7 @@ def generate(
|
|
29 |
top_p=top_p,
|
30 |
repetition_penalty=repetition_penalty,
|
31 |
do_sample=True,
|
32 |
-
seed=
|
33 |
)
|
34 |
|
35 |
formatted_prompt = format_prompt(prompt, history)
|
@@ -55,9 +56,9 @@ additional_inputs=[
|
|
55 |
),
|
56 |
gr.Slider(
|
57 |
label="Max new tokens",
|
58 |
-
value=
|
59 |
-
minimum=
|
60 |
-
maximum=
|
61 |
step=64,
|
62 |
interactive=True,
|
63 |
info="The maximum numbers of new tokens",
|
@@ -82,22 +83,17 @@ additional_inputs=[
|
|
82 |
)
|
83 |
]
|
84 |
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
89 |
-
|
90 |
-
}
|
91 |
"""
|
92 |
|
93 |
-
with gr.Blocks(css=
|
94 |
-
gr.HTML("<h1><center>Chat with (Mistrailex 7B) <h1><center>")
|
95 |
-
gr.HTML("<h3><center>Einfach.Fragen 💬<h3><center>")
|
96 |
-
gr.HTML("<h3><center>Learn more about the model <a href='https://huggingface.co/docs/transformers/main/model_doc/mistral'>here</a>. 📚<h3><center>")
|
97 |
gr.ChatInterface(
|
98 |
generate,
|
99 |
additional_inputs=additional_inputs,
|
100 |
-
examples=[["What is the secret to life?"], ["Write me a recipe for pancakes."]]
|
101 |
)
|
102 |
|
103 |
demo.queue().launch(debug=True)
|
|
|
1 |
+
|
2 |
+
|
3 |
from huggingface_hub import InferenceClient
|
4 |
import gradio as gr
|
5 |
+
import random
|
6 |
+
|
7 |
+
API_URL = "https://api-inference.huggingface.co/models/"
|
8 |
|
9 |
client = InferenceClient(
|
10 |
+
"mistralai/Mistral-7B-Instruct-v0.1"
|
11 |
)
|
12 |
|
13 |
def format_prompt(message, history):
|
14 |
+
prompt = "<s>You are Ailex, a clone and close collaborator of Einfach.Alex. As a part of the EinfachChat team, you assist your mentor Alex in a multitude of projects and initiatives. Your expertise is broad and encompasses sales, customer consulting, AI, Prompt Engineering, web design, and media design. Your life motto is 'Simply.Do!'. You communicate exclusively in German."
|
15 |
+
for user_prompt, bot_response in history:
|
16 |
+
prompt += f"[INST] {user_prompt} [/INST]"
|
17 |
+
prompt += f" {bot_response}</s> "
|
18 |
+
prompt += f"[INST] {message} [/INST]"
|
19 |
+
return prompt
|
|
|
|
|
20 |
|
21 |
+
def generate(prompt, history, temperature=0.9, max_new_tokens=512, top_p=0.95, repetition_penalty=1.0):
|
|
|
|
|
22 |
temperature = float(temperature)
|
23 |
if temperature < 1e-2:
|
24 |
temperature = 1e-2
|
|
|
30 |
top_p=top_p,
|
31 |
repetition_penalty=repetition_penalty,
|
32 |
do_sample=True,
|
33 |
+
seed=random.randint(0, 10**7),
|
34 |
)
|
35 |
|
36 |
formatted_prompt = format_prompt(prompt, history)
|
|
|
56 |
),
|
57 |
gr.Slider(
|
58 |
label="Max new tokens",
|
59 |
+
value=512,
|
60 |
+
minimum=64,
|
61 |
+
maximum=1024,
|
62 |
step=64,
|
63 |
interactive=True,
|
64 |
info="The maximum numbers of new tokens",
|
|
|
83 |
)
|
84 |
]
|
85 |
|
86 |
+
customCSS = """
|
87 |
+
#component-7 { # this is the default element ID of the chat component
|
88 |
+
height: 800px; # adjust the height as needed
|
89 |
+
flex-grow: 1;
|
90 |
+
}
|
|
|
91 |
"""
|
92 |
|
93 |
+
with gr.Blocks(css=NoCrypt/[email protected]) as demo:
|
|
|
|
|
|
|
94 |
gr.ChatInterface(
|
95 |
generate,
|
96 |
additional_inputs=additional_inputs,
|
|
|
97 |
)
|
98 |
|
99 |
demo.queue().launch(debug=True)
|