jonaschua commited on
Commit
5dfffe1
·
verified ·
1 Parent(s): 0a93e35

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -2
app.py CHANGED
@@ -27,9 +27,15 @@ def choose_model(model_name):
27
  elif model_name == "Llama3.1-8b-Instruct":
28
  model = "meta-llama/Llama-3.1-8B-Instruct"
29
 
30
- elif model_name == "Llama2-13b-chat"
31
  model = "meta-llama/Llama-2-13b-chat-hf"
32
 
 
 
 
 
 
 
33
  client = InferenceClient(model, token=os.getenv('deepseekv2'))
34
  return client
35
 
@@ -68,7 +74,7 @@ demo = gr.ChatInterface(
68
  respond,
69
 
70
  additional_inputs=[
71
- gr.Dropdown(["Qwen1.5", "Llama2-13b-chat", "Llama3-8b-Instruct", "Llama3.1-8b-Instruct"], label="Select Model"),
72
  gr.Textbox(value="You are a friendly and helpful Chatbot, be concise and straight to the point, avoid excessive reasoning.", label="System message"),
73
  gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
74
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
 
27
  elif model_name == "Llama3.1-8b-Instruct":
28
  model = "meta-llama/Llama-3.1-8B-Instruct"
29
 
30
+ elif model_name == "Llama2-13b-chat":
31
  model = "meta-llama/Llama-2-13b-chat-hf"
32
 
33
+ elif model_name == "Gemma-2-2b":
34
+ model = "google/gemma-2-2b-it"
35
+
36
+ elif model_name == "Mixtral-8x7B-Instruct":
37
+ model = "mistralai/Mixtral-8x7B-Instruct-v0.1"
38
+
39
  client = InferenceClient(model, token=os.getenv('deepseekv2'))
40
  return client
41
 
 
74
  respond,
75
 
76
  additional_inputs=[
77
+ gr.Dropdown(["Qwen1.5", "Gemma-2-2b", "Llama2-13b-chat", "Llama3-8b-Instruct", "Llama3.1-8b-Instruct", "Mixtral-8x7B-Instruct"], label="Select Model"),
78
  gr.Textbox(value="You are a friendly and helpful Chatbot, be concise and straight to the point, avoid excessive reasoning.", label="System message"),
79
  gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
80
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),