kamran-r123 commited on
Commit
bab92d5
·
verified ·
1 Parent(s): f79168b

Update main.py

Browse files
Files changed (1) hide show
  1. main.py +6 -5
main.py CHANGED
@@ -4,12 +4,12 @@ from llama_cpp import Llama
4
  import uvicorn
5
  import prompt_style
6
  import os
 
7
 
8
- model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3-GGUF"
9
- model = Llama.from_pretrained(repo_id=model_id, filename="*-v3_q6.gguf", n_gpu_layers=-1, n_ctx=4096, verbose=False)
10
 
11
- # model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3"
12
- # client = InferenceClient(token=os.getenv('HF_TOKEN'), model=model_id)
 
13
 
14
  class Item(BaseModel):
15
  prompt: str
@@ -51,4 +51,5 @@ async def generate_text(item: Item):
51
 
52
  @app.get("/")
53
  def read_root():
54
- return {"Hello": "World!"}
 
 
4
  import uvicorn
5
  import prompt_style
6
  import os
7
+ from huggingface_hub import hf_hub_download
8
 
 
 
9
 
10
+ model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3-GGUF"
11
+ model_path = hf_hub_download(repo_id=model_id, filename="*-v3_q6.gguf", token=os.environment["HF_TOKEN"])
12
+ model = Llama(model_path=model_path, n_gpu_layers=-1, n_ctx=4096, verbose=False)
13
 
14
  class Item(BaseModel):
15
  prompt: str
 
51
 
52
  @app.get("/")
53
  def read_root():
54
+
55
+ return {"Hello": os.environment["HF_TOKEN"]}