Update main.py
Browse files
main.py
CHANGED
@@ -4,12 +4,12 @@ from llama_cpp import Llama
|
|
4 |
import uvicorn
|
5 |
import prompt_style
|
6 |
import os
|
|
|
7 |
|
8 |
-
model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3-GGUF"
|
9 |
-
model = Llama.from_pretrained(repo_id=model_id, filename="*-v3_q6.gguf", n_gpu_layers=-1, n_ctx=4096, verbose=False)
|
10 |
|
11 |
-
|
12 |
-
|
|
|
13 |
|
14 |
class Item(BaseModel):
|
15 |
prompt: str
|
@@ -51,4 +51,5 @@ async def generate_text(item: Item):
|
|
51 |
|
52 |
@app.get("/")
|
53 |
def read_root():
|
54 |
-
|
|
|
|
4 |
import uvicorn
|
5 |
import prompt_style
|
6 |
import os
|
7 |
+
from huggingface_hub import hf_hub_download
|
8 |
|
|
|
|
|
9 |
|
10 |
+
model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3-GGUF"
|
11 |
+
model_path = hf_hub_download(repo_id=model_id, filename="*-v3_q6.gguf", token=os.environment["HF_TOKEN"])
|
12 |
+
model = Llama(model_path=model_path, n_gpu_layers=-1, n_ctx=4096, verbose=False)
|
13 |
|
14 |
class Item(BaseModel):
|
15 |
prompt: str
|
|
|
51 |
|
52 |
@app.get("/")
|
53 |
def read_root():
|
54 |
+
|
55 |
+
return {"Hello": os.environment["HF_TOKEN"]}
|