pabloce commited on
Commit
ac7a09d
·
verified ·
1 Parent(s): 43af2cd

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -28,7 +28,7 @@ def respond(
28
  from llama_cpp_agent.chat_history.messages import Roles
29
  print(model)
30
  llm = Llama(
31
- model_path="models/Meta-Llama-3-70B-Instruct-Q3_K_M.gguf",
32
  n_gpu_layers=81,
33
  )
34
  provider = LlamaCppPythonProvider(llm)
@@ -82,7 +82,7 @@ demo = gr.ChatInterface(
82
  step=0.05,
83
  label="Top-p (nucleus sampling)",
84
  ),
85
- gr.Dropdown(['Meta-Llama-3-70B-Instruct-Q3_K_M.gguf', 'Smaug-Llama-3-70B-Instruct-Q3_K_M.gguf'], label="Model"),
86
  ],
87
  theme=gr.themes.Soft(primary_hue="green", secondary_hue="indigo", neutral_hue="zinc",font=[gr.themes.GoogleFont("Exo 2"), "ui-sans-serif", "system-ui", "sans-serif"]).set(
88
  block_background_fill_dark="*neutral_800"
 
28
  from llama_cpp_agent.chat_history.messages import Roles
29
  print(model)
30
  llm = Llama(
31
+ model_path=f"models/{model}",
32
  n_gpu_layers=81,
33
  )
34
  provider = LlamaCppPythonProvider(llm)
 
82
  step=0.05,
83
  label="Top-p (nucleus sampling)",
84
  ),
85
+ gr.Dropdown(['Meta-Llama-3-70B-Instruct-Q3_K_M.gguf', 'Smaug-Llama-3-70B-Instruct-Q3_K_M.gguf'], value="Meta-Llama-3-70B-Instruct-Q3_K_M.gguf", label="Model"),
86
  ],
87
  theme=gr.themes.Soft(primary_hue="green", secondary_hue="indigo", neutral_hue="zinc",font=[gr.themes.GoogleFont("Exo 2"), "ui-sans-serif", "system-ui", "sans-serif"]).set(
88
  block_background_fill_dark="*neutral_800"