Luigi commited on
Commit
4d633ef
·
1 Parent(s): 8287454

switch to 1.5b

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -3,8 +3,8 @@ from llama_cpp import Llama
3
  from huggingface_hub import hf_hub_download
4
 
5
  hf_hub_download(
6
- repo_id="Qwen/Qwen2.5-3B-Instruct-GGUF",
7
- filename="qwen2.5-3b-instruct-q4_k_m.gguf",
8
  local_dir="./models",
9
  )
10
 
@@ -12,7 +12,7 @@ hf_hub_download(
12
  @st.cache_resource
13
  def load_model():
14
  return Llama(
15
- model_path="models/qwen2.5-3b-instruct-q4_k_m.gguf",
16
  n_ctx=2048,
17
  n_threads=6,
18
  n_batch=8,
@@ -28,7 +28,7 @@ llm = load_model()
28
  if "chat_history" not in st.session_state:
29
  st.session_state.chat_history = []
30
 
31
- st.title("🧠 Qwen2.5-3B-Instruct (Streamlit + GGUF)")
32
  st.caption("Powered by `llama.cpp` and `llama-cpp-python` | 4-bit Q4_K_M inference")
33
 
34
  with st.sidebar:
 
3
  from huggingface_hub import hf_hub_download
4
 
5
  hf_hub_download(
6
+ repo_id="Qwen/Qwen2.5-1.5B-Instruct-GGUF",
7
+ filename="qwen2.5-1.5b-instruct-q4_k_m.gguf",
8
  local_dir="./models",
9
  )
10
 
 
12
  @st.cache_resource
13
  def load_model():
14
  return Llama(
15
+ model_path="models/qwen2.5-1.5b-instruct-q4_k_m.gguf",
16
  n_ctx=2048,
17
  n_threads=6,
18
  n_batch=8,
 
28
  if "chat_history" not in st.session_state:
29
  st.session_state.chat_history = []
30
 
31
+ st.title("🧠 Qwen2.5-1.5B-Instruct (Streamlit + GGUF)")
32
  st.caption("Powered by `llama.cpp` and `llama-cpp-python` | 4-bit Q4_K_M inference")
33
 
34
  with st.sidebar: