Update app/llm.py
Browse files- app/llm.py +2 -2
    	
        app/llm.py
    CHANGED
    
    | @@ -31,9 +31,9 @@ class ChatModel(BaseModel): | |
| 31 | 
             
            llm_chat = llama_cpp.Llama.from_pretrained(
         | 
| 32 | 
             
                repo_id="Qwen/Qwen1.5-0.5B-Chat-GGUF",
         | 
| 33 | 
             
                filename="*q4_0.gguf",
         | 
| 34 | 
            -
                 | 
| 35 | 
             
                verbose=False,
         | 
| 36 | 
            -
                 n_ctx= | 
| 37 | 
             
                 n_gpu_layers=0,
         | 
| 38 | 
             
                #chat_format="llama-2"
         | 
| 39 | 
             
            )
         | 
|  | |
| 31 | 
             
            llm_chat = llama_cpp.Llama.from_pretrained(
         | 
| 32 | 
             
                repo_id="Qwen/Qwen1.5-0.5B-Chat-GGUF",
         | 
| 33 | 
             
                filename="*q4_0.gguf",
         | 
| 34 | 
            +
                tokenizer=llama_cpp.llama_tokenizer.LlamaHFTokenizer.from_pretrained("Qwen/Qwen1.5-0.5B-Chat"),
         | 
| 35 | 
             
                verbose=False,
         | 
| 36 | 
            +
                 n_ctx=128,
         | 
| 37 | 
             
                 n_gpu_layers=0,
         | 
| 38 | 
             
                #chat_format="llama-2"
         | 
| 39 | 
             
            )
         |