pabloce commited on
Commit
53e74c5
·
verified ·
1 Parent(s): cd6bea7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -7,7 +7,7 @@ from huggingface_hub import hf_hub_download
7
  subprocess.run('pip install llama-cpp-python==0.2.75 --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu124', shell=True)
8
  subprocess.run('pip install llama-cpp-agent==0.2.10', shell=True)
9
 
10
- hf_hub_download(repo_id="TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF", filename="mixtral-8x7b-instruct-v0.1.Q5_K_M.gguf", local_dir = "./models")
11
 
12
  @spaces.GPU(duration=120)
13
  def respond(
@@ -26,7 +26,7 @@ def respond(
26
  from llama_cpp_agent.chat_history.messages import Roles
27
 
28
  llm = Llama(
29
- model_path="models/mixtral-8x7b-instruct-v0.1.Q5_K_M.gguf",
30
  n_gpu_layers=33,
31
  )
32
  provider = LlamaCppPythonProvider(llm)
 
7
  subprocess.run('pip install llama-cpp-python==0.2.75 --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu124', shell=True)
8
  subprocess.run('pip install llama-cpp-agent==0.2.10', shell=True)
9
 
10
+ hf_hub_download(repo_id="TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF", filename="mixtral-8x7b-instruct-v0.1.Q6_K.gguf", local_dir = "./models")
11
 
12
  @spaces.GPU(duration=120)
13
  def respond(
 
26
  from llama_cpp_agent.chat_history.messages import Roles
27
 
28
  llm = Llama(
29
+ model_path="models/mixtral-8x7b-instruct-v0.1.Q6_K.gguf",
30
  n_gpu_layers=33,
31
  )
32
  provider = LlamaCppPythonProvider(llm)