Spaces:
Running
Running
Commit
·
00f6950
1
Parent(s):
d79eb5f
Change model to Llama
Browse files
app.py
CHANGED
@@ -11,9 +11,12 @@ from langchain.schema import Document
|
|
11 |
load_dotenv()
|
12 |
hftoken = os.getenv("HFTOKEN")
|
13 |
|
|
|
|
|
|
|
14 |
login(token=hftoken)
|
15 |
|
16 |
-
client = InferenceClient("
|
17 |
embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2")
|
18 |
|
19 |
vector_store = None
|
@@ -76,7 +79,7 @@ def chatbot_response(message, history, system_message, max_tokens, temperature,
|
|
76 |
yield status
|
77 |
|
78 |
context = query_faiss(message) if vector_store else None
|
79 |
-
messages = [{"role": "system", "content": system_message
|
80 |
|
81 |
for val in history:
|
82 |
if val[0]: messages.append({"role": "user", "content": val[0]})
|
@@ -120,7 +123,7 @@ with gr.Blocks(theme=gr.themes.Soft()) as demo:
|
|
120 |
with gr.Column():
|
121 |
gr.Markdown("## ⚙️ Paramètres")
|
122 |
with gr.Accordion(" Paramètres avancés", open=False):
|
123 |
-
system_message = gr.Textbox(value=
|
124 |
max_tokens = gr.Slider(1, 2048, value=800, step=1, label="Max tokens")
|
125 |
temperature = gr.Slider(0.1, 4.0, value=0.3, step=0.1, label="Température")
|
126 |
top_p = gr.Slider(0.1, 1.0, value=0.95, step=0.05, label="Top-p")
|
|
|
11 |
load_dotenv()
|
12 |
hftoken = os.getenv("HFTOKEN")
|
13 |
|
14 |
+
prompt_systeme = os.getenv("PROMPT_SYSTEM")
|
15 |
+
|
16 |
+
|
17 |
login(token=hftoken)
|
18 |
|
19 |
+
client = InferenceClient(model="meta-llama/Llama-3.3-70B-Instruct", token=hftoken)
|
20 |
embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2")
|
21 |
|
22 |
vector_store = None
|
|
|
79 |
yield status
|
80 |
|
81 |
context = query_faiss(message) if vector_store else None
|
82 |
+
messages = [{"role": "system", "content": system_message}]
|
83 |
|
84 |
for val in history:
|
85 |
if val[0]: messages.append({"role": "user", "content": val[0]})
|
|
|
123 |
with gr.Column():
|
124 |
gr.Markdown("## ⚙️ Paramètres")
|
125 |
with gr.Accordion(" Paramètres avancés", open=False):
|
126 |
+
system_message = gr.Textbox(value=prompt_systeme, label="Message système")
|
127 |
max_tokens = gr.Slider(1, 2048, value=800, step=1, label="Max tokens")
|
128 |
temperature = gr.Slider(0.1, 4.0, value=0.3, step=0.1, label="Température")
|
129 |
top_p = gr.Slider(0.1, 1.0, value=0.95, step=0.05, label="Top-p")
|