Spaces:
				
			
			
	
			
			
					
		Running
		
	
	
	
			
			
	
	
	
	
		
		
					
		Running
		
	Update app.py
Browse files
    	
        app.py
    CHANGED
    
    | 
         @@ -267,28 +267,28 @@ def simple_chat(message: dict, temperature: float = 0.8, max_length: int = 4096, 
     | 
|
| 267 | 
         
             
                        eos_token_id=[151329, 151336, 151338],
         
     | 
| 268 | 
         
             
                    )
         
     | 
| 269 | 
         | 
| 270 | 
         
            -
                    gen_kwargs = {**input_ids, **generate_kwargs}
         
     | 
| 271 | 
         
            -
             
     | 
| 272 | 
         
            -
                    with torch.no_grad():
         
     | 
| 273 | 
         
            -
                        thread = Thread(target=model.generate, kwargs=gen_kwargs)
         
     | 
| 274 | 
         
            -
                        thread.start()
         
     | 
| 275 | 
         
            -
                        buffer = ""
         
     | 
| 276 | 
         
            -
                        for new_text in streamer:
         
     | 
| 277 | 
         
            -
                            buffer += new_text
         
     | 
| 278 | 
         
            -
                            yield buffer
         
     | 
| 279 | 
         
            -
             
     | 
| 280 | 
         
            -
                    print("--------------")
         
     | 
| 281 | 
         
            -
                    print("Buffer: ")
         
     | 
| 282 | 
         
            -
                    print("  ")
         
     | 
| 283 | 
         
            -
                    print(buffer)
         
     | 
| 284 | 
         
            -
                    print("  ")
         
     | 
| 285 | 
         
            -
                    print("--------------")
         
     | 
| 286 | 
         | 
| 287 | 
         
             
                    # with torch.no_grad():
         
     | 
| 288 | 
         
            -
                    #      
     | 
| 289 | 
         
            -
                    #      
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 290 | 
         | 
| 291 | 
         
            -
                    generated_text = buffer
         
     | 
| 292 | 
         | 
| 293 | 
         | 
| 294 | 
         
             
                    return PlainTextResponse(generated_text)
         
     | 
| 
         | 
|
| 267 | 
         
             
                        eos_token_id=[151329, 151336, 151338],
         
     | 
| 268 | 
         
             
                    )
         
     | 
| 269 | 
         | 
| 270 | 
         
            +
                    # gen_kwargs = {**input_ids, **generate_kwargs}
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 271 | 
         | 
| 272 | 
         
             
                    # with torch.no_grad():
         
     | 
| 273 | 
         
            +
                    #     thread = Thread(target=model.generate, kwargs=gen_kwargs)
         
     | 
| 274 | 
         
            +
                    #     thread.start()
         
     | 
| 275 | 
         
            +
                    #     buffer = ""
         
     | 
| 276 | 
         
            +
                    #     for new_text in streamer:
         
     | 
| 277 | 
         
            +
                    #         buffer += new_text
         
     | 
| 278 | 
         
            +
                    #         yield buffer
         
     | 
| 279 | 
         
            +
             
     | 
| 280 | 
         
            +
                    # print("--------------")
         
     | 
| 281 | 
         
            +
                    # print("Buffer: ")
         
     | 
| 282 | 
         
            +
                    # print("  ")
         
     | 
| 283 | 
         
            +
                    # print(buffer)
         
     | 
| 284 | 
         
            +
                    # print("  ")
         
     | 
| 285 | 
         
            +
                    # print("--------------")
         
     | 
| 286 | 
         
            +
             
     | 
| 287 | 
         
            +
                    with torch.no_grad():
         
     | 
| 288 | 
         
            +
                        generated_ids = model.generate(input_ids['input_ids'], **generate_kwargs)
         
     | 
| 289 | 
         
            +
                        generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=True)
         
     | 
| 290 | 
         | 
| 291 | 
         
            +
                    #generated_text = buffer
         
     | 
| 292 | 
         | 
| 293 | 
         | 
| 294 | 
         
             
                    return PlainTextResponse(generated_text)
         
     |