ionosphere commited on
Commit
00f6950
·
1 Parent(s): d79eb5f

Change model to Llama

Browse files
Files changed (1) hide show
  1. app.py +6 -3
app.py CHANGED
@@ -11,9 +11,12 @@ from langchain.schema import Document
11
  load_dotenv()
12
  hftoken = os.getenv("HFTOKEN")
13
 
 
 
 
14
  login(token=hftoken)
15
 
16
- client = InferenceClient("HuggingFaceH4/zephyr-7b-beta", token=hftoken)
17
  embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2")
18
 
19
  vector_store = None
@@ -76,7 +79,7 @@ def chatbot_response(message, history, system_message, max_tokens, temperature,
76
  yield status
77
 
78
  context = query_faiss(message) if vector_store else None
79
- messages = [{"role": "system", "content": system_message + " Réponds uniquement en français."}]
80
 
81
  for val in history:
82
  if val[0]: messages.append({"role": "user", "content": val[0]})
@@ -120,7 +123,7 @@ with gr.Blocks(theme=gr.themes.Soft()) as demo:
120
  with gr.Column():
121
  gr.Markdown("## ⚙️ Paramètres")
122
  with gr.Accordion(" Paramètres avancés", open=False):
123
- system_message = gr.Textbox(value="Réponds de façon simple et claire.", label="Message système")
124
  max_tokens = gr.Slider(1, 2048, value=800, step=1, label="Max tokens")
125
  temperature = gr.Slider(0.1, 4.0, value=0.3, step=0.1, label="Température")
126
  top_p = gr.Slider(0.1, 1.0, value=0.95, step=0.05, label="Top-p")
 
11
  load_dotenv()
12
  hftoken = os.getenv("HFTOKEN")
13
 
14
+ prompt_systeme = os.getenv("PROMPT_SYSTEM")
15
+
16
+
17
  login(token=hftoken)
18
 
19
+ client = InferenceClient(model="meta-llama/Llama-3.3-70B-Instruct", token=hftoken)
20
  embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2")
21
 
22
  vector_store = None
 
79
  yield status
80
 
81
  context = query_faiss(message) if vector_store else None
82
+ messages = [{"role": "system", "content": system_message}]
83
 
84
  for val in history:
85
  if val[0]: messages.append({"role": "user", "content": val[0]})
 
123
  with gr.Column():
124
  gr.Markdown("## ⚙️ Paramètres")
125
  with gr.Accordion(" Paramètres avancés", open=False):
126
+ system_message = gr.Textbox(value=prompt_systeme, label="Message système")
127
  max_tokens = gr.Slider(1, 2048, value=800, step=1, label="Max tokens")
128
  temperature = gr.Slider(0.1, 4.0, value=0.3, step=0.1, label="Température")
129
  top_p = gr.Slider(0.1, 1.0, value=0.95, step=0.05, label="Top-p")