Spaces:
				
			
			
	
			
			
		Runtime error
		
	
	
	
			
			
	
	
	
	
		
		
		Runtime error
		
	added model download
Browse files
    	
        app.py
    CHANGED
    
    | @@ -19,10 +19,11 @@ def download_model(): | |
| 19 | 
             
                print(f'Downloading model {REPO_ID}/{FILENAME}')
         | 
| 20 | 
             
                m = hf_hub_download(repo_id=REPO_ID, filename=FILENAME)
         | 
| 21 | 
             
                print(f'status: {m}')
         | 
|  | |
| 22 |  | 
| 23 | 
            -
            def load_model():
         | 
| 24 | 
             
                from llama_cpp import Llama, LlamaGrammar
         | 
| 25 | 
            -
                model_file= | 
| 26 | 
             
                llm = Llama(
         | 
| 27 | 
             
                    model_path=model_file,
         | 
| 28 | 
             
                    n_gpu_layers=-1, verbose=False
         | 
| @@ -58,8 +59,8 @@ def load_model(): | |
| 58 | 
             
                    print(output['choices'][0]['text'])
         | 
| 59 | 
             
                    print()
         | 
| 60 |  | 
| 61 | 
            -
             | 
| 62 | 
            -
            load_model()
         | 
| 63 | 
             
            demo = gr.Interface(fn=greet, inputs=gr.Number(), outputs=gr.Text())
         | 
| 64 | 
             
            demo.launch(share=False)
         | 
| 65 |  | 
|  | |
| 19 | 
             
                print(f'Downloading model {REPO_ID}/{FILENAME}')
         | 
| 20 | 
             
                m = hf_hub_download(repo_id=REPO_ID, filename=FILENAME)
         | 
| 21 | 
             
                print(f'status: {m}')
         | 
| 22 | 
            +
                return m
         | 
| 23 |  | 
| 24 | 
            +
            def load_model(fp):
         | 
| 25 | 
             
                from llama_cpp import Llama, LlamaGrammar
         | 
| 26 | 
            +
                model_file=fp
         | 
| 27 | 
             
                llm = Llama(
         | 
| 28 | 
             
                    model_path=model_file,
         | 
| 29 | 
             
                    n_gpu_layers=-1, verbose=False
         | 
|  | |
| 59 | 
             
                    print(output['choices'][0]['text'])
         | 
| 60 | 
             
                    print()
         | 
| 61 |  | 
| 62 | 
            +
             | 
| 63 | 
            +
            load_model(download_model())
         | 
| 64 | 
             
            demo = gr.Interface(fn=greet, inputs=gr.Number(), outputs=gr.Text())
         | 
| 65 | 
             
            demo.launch(share=False)
         | 
| 66 |  | 
