Christoph Holthaus
commited on
Commit
·
ce6efbb
1
Parent(s):
9669900
update texts
Browse files
app.py
CHANGED
@@ -12,8 +12,8 @@ llama_model_name = "TheBloke/dolphin-2.2.1-AshhLimaRP-Mistral-7B-GGUF"
|
|
12 |
print("! INITING DONE !")
|
13 |
|
14 |
# Preparing things to work
|
15 |
-
title = "Demo for 7B Models - Quantized"
|
16 |
-
|
17 |
Quantized to run in the free tier hosting.
|
18 |
Have a quick way to test models or share them with others without hassle.
|
19 |
It runs slow, as it's on cpu. Usable for basic tests.
|
@@ -35,7 +35,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStream
|
|
35 |
DESCRIPTION = "# Mistral-7B"
|
36 |
|
37 |
if torch.cuda.is_available():
|
38 |
-
DESCRIPTION += "\n<p>This space is
|
39 |
|
40 |
MAX_MAX_NEW_TOKENS = 2048
|
41 |
DEFAULT_MAX_NEW_TOKENS = 1024
|
@@ -142,7 +142,8 @@ chat_interface = gr.ChatInterface(
|
|
142 |
)
|
143 |
|
144 |
with gr.Blocks(css="style.css") as demo:
|
145 |
-
gr.Markdown(
|
|
|
146 |
gr.DuplicateButton(
|
147 |
value="Duplicate Space for private use",
|
148 |
elem_id="duplicate-button",
|
|
|
12 |
print("! INITING DONE !")
|
13 |
|
14 |
# Preparing things to work
|
15 |
+
title = "# Demo for 7B Models - Quantized"
|
16 |
+
descr = '''<h1>Demo for 7B Models - Quantized</h1>
|
17 |
Quantized to run in the free tier hosting.
|
18 |
Have a quick way to test models or share them with others without hassle.
|
19 |
It runs slow, as it's on cpu. Usable for basic tests.
|
|
|
35 |
DESCRIPTION = "# Mistral-7B"
|
36 |
|
37 |
if torch.cuda.is_available():
|
38 |
+
DESCRIPTION += "\n<p>This space is using CPU only. Use a different one if you want to go fast and use GPU. </p>"
|
39 |
|
40 |
MAX_MAX_NEW_TOKENS = 2048
|
41 |
DEFAULT_MAX_NEW_TOKENS = 1024
|
|
|
142 |
)
|
143 |
|
144 |
with gr.Blocks(css="style.css") as demo:
|
145 |
+
gr.Markdown(title)
|
146 |
+
gr.Markdown(descr)
|
147 |
gr.DuplicateButton(
|
148 |
value="Duplicate Space for private use",
|
149 |
elem_id="duplicate-button",
|