Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -9,8 +9,8 @@ from huggingface_hub import hf_hub_download
|
|
9 |
|
10 |
llm = Llama(
|
11 |
model_path=hf_hub_download(
|
12 |
-
repo_id=os.environ.get("REPO_ID", "
|
13 |
-
filename=os.environ.get("MODEL_FILE", "
|
14 |
),
|
15 |
n_ctx=2048,
|
16 |
n_gpu_layers=50, # change n_gpu_layers if you have more or less VRAM
|
@@ -60,13 +60,12 @@ def generate_text(message, history):
|
|
60 |
|
61 |
demo = gr.ChatInterface(
|
62 |
generate_text,
|
63 |
-
title="
|
64 |
-
description="
|
65 |
-
examples=["tell me everything about
|
66 |
cache_examples=True,
|
67 |
retry_btn=None,
|
68 |
undo_btn="Delete Previous",
|
69 |
clear_btn="Clear",
|
70 |
-
|
71 |
)
|
72 |
demo.launch(max_threads=10)
|
|
|
9 |
|
10 |
llm = Llama(
|
11 |
model_path=hf_hub_download(
|
12 |
+
repo_id=os.environ.get("REPO_ID", "Severian/ANIMA-Phi-Neptune-Mistral-7B-gguf"),
|
13 |
+
filename=os.environ.get("MODEL_FILE", "anima-phi-neptune-mistral-7b.Q4_K_M.gguf"),
|
14 |
),
|
15 |
n_ctx=2048,
|
16 |
n_gpu_layers=50, # change n_gpu_layers if you have more or less VRAM
|
|
|
60 |
|
61 |
demo = gr.ChatInterface(
|
62 |
generate_text,
|
63 |
+
title="A N I M A",
|
64 |
+
description="ANIMA is an expert in various scientific disciplines.",
|
65 |
+
examples=["tell me everything about biomimicry"],
|
66 |
cache_examples=True,
|
67 |
retry_btn=None,
|
68 |
undo_btn="Delete Previous",
|
69 |
clear_btn="Clear",
|
|
|
70 |
)
|
71 |
demo.launch(max_threads=10)
|