Update main.py
Browse files
    	
        main.py
    CHANGED
    
    | @@ -4,12 +4,12 @@ from llama_cpp import Llama | |
| 4 | 
             
            import uvicorn
         | 
| 5 | 
             
            import prompt_style
         | 
| 6 | 
             
            import os
         | 
|  | |
| 7 |  | 
| 8 | 
            -
            model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3-GGUF"
         | 
| 9 | 
            -
            model = Llama.from_pretrained(repo_id=model_id, filename="*-v3_q6.gguf", n_gpu_layers=-1, n_ctx=4096, verbose=False)
         | 
| 10 |  | 
| 11 | 
            -
             | 
| 12 | 
            -
             | 
|  | |
| 13 |  | 
| 14 | 
             
            class Item(BaseModel):
         | 
| 15 | 
             
                prompt: str
         | 
| @@ -51,4 +51,5 @@ async def generate_text(item: Item): | |
| 51 |  | 
| 52 | 
             
            @app.get("/")
         | 
| 53 | 
             
            def read_root():
         | 
| 54 | 
            -
                 | 
|  | 
|  | |
| 4 | 
             
            import uvicorn
         | 
| 5 | 
             
            import prompt_style
         | 
| 6 | 
             
            import os
         | 
| 7 | 
            +
            from huggingface_hub import hf_hub_download
         | 
| 8 |  | 
|  | |
|  | |
| 9 |  | 
| 10 | 
            +
            model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3-GGUF"
         | 
| 11 | 
            +
            model_path = hf_hub_download(repo_id=model_id, filename="*-v3_q6.gguf", token=os.environment["HF_TOKEN"])
         | 
| 12 | 
            +
            model = Llama(model_path=model_path, n_gpu_layers=-1, n_ctx=4096, verbose=False)
         | 
| 13 |  | 
| 14 | 
             
            class Item(BaseModel):
         | 
| 15 | 
             
                prompt: str
         | 
|  | |
| 51 |  | 
| 52 | 
             
            @app.get("/")
         | 
| 53 | 
             
            def read_root():
         | 
| 54 | 
            +
                
         | 
| 55 | 
            +
                return {"Hello": os.environment["HF_TOKEN"]}
         |