WesanCZE commited on
Commit
f1e5dd5
·
verified ·
1 Parent(s): e7bbf43

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -3
app.py CHANGED
@@ -2,15 +2,21 @@ import os
2
  import torch
3
  from llama_cpp import Llama
4
  import gradio as gr
 
5
 
6
- # 🔹 Stáhneme model, pokud není dostupný
 
 
 
 
 
 
7
  MODEL_NAME = "TheBloke/Mistral-7B-Instruct-GGUF"
8
  MODEL_FILE = "Mistral-7B-Instruct-Q4_K_M.gguf"
9
  MODEL_PATH = f"./{MODEL_FILE}"
10
 
11
  if not os.path.exists(MODEL_PATH):
12
- from huggingface_hub import hf_hub_download
13
- MODEL_PATH = hf_hub_download(repo_id=MODEL_NAME, filename=MODEL_FILE)
14
 
15
  # 🔹 Načteme model do RAM
16
  llm = Llama(model_path=MODEL_PATH, n_ctx=1024, n_threads=8, verbose=False)
 
2
  import torch
3
  from llama_cpp import Llama
4
  import gradio as gr
5
+ from huggingface_hub import login, hf_hub_download
6
 
7
+ # 🔹 Načteme API token bezpečně ze "Secrets"
8
+ HF_TOKEN = os.getenv("HF_TOKEN")
9
+
10
+ # Přihlásíme se k Hugging Face API
11
+ login(HF_TOKEN)
12
+
13
+ # 🔹 Stáhneme model
14
  MODEL_NAME = "TheBloke/Mistral-7B-Instruct-GGUF"
15
  MODEL_FILE = "Mistral-7B-Instruct-Q4_K_M.gguf"
16
  MODEL_PATH = f"./{MODEL_FILE}"
17
 
18
  if not os.path.exists(MODEL_PATH):
19
+ MODEL_PATH = hf_hub_download(repo_id=MODEL_NAME, filename=MODEL_FILE, token=HF_TOKEN)
 
20
 
21
  # 🔹 Načteme model do RAM
22
  llm = Llama(model_path=MODEL_PATH, n_ctx=1024, n_threads=8, verbose=False)