mateoluksenberg commited on
Commit
c002c58
·
verified ·
1 Parent(s): 6e8411f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +19 -19
app.py CHANGED
@@ -267,28 +267,28 @@ def simple_chat(message: dict, temperature: float = 0.8, max_length: int = 4096,
267
  eos_token_id=[151329, 151336, 151338],
268
  )
269
 
270
- gen_kwargs = {**input_ids, **generate_kwargs}
271
-
272
- with torch.no_grad():
273
- thread = Thread(target=model.generate, kwargs=gen_kwargs)
274
- thread.start()
275
- buffer = ""
276
- for new_text in streamer:
277
- buffer += new_text
278
- yield buffer
279
-
280
- print("--------------")
281
- print("Buffer: ")
282
- print(" ")
283
- print(buffer)
284
- print(" ")
285
- print("--------------")
286
 
287
  # with torch.no_grad():
288
- # generated_ids = model.generate(input_ids['input_ids'], **generate_kwargs)
289
- # generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=True)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
290
 
291
- generated_text = buffer
292
 
293
 
294
  return PlainTextResponse(generated_text)
 
267
  eos_token_id=[151329, 151336, 151338],
268
  )
269
 
270
+ # gen_kwargs = {**input_ids, **generate_kwargs}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
271
 
272
  # with torch.no_grad():
273
+ # thread = Thread(target=model.generate, kwargs=gen_kwargs)
274
+ # thread.start()
275
+ # buffer = ""
276
+ # for new_text in streamer:
277
+ # buffer += new_text
278
+ # yield buffer
279
+
280
+ # print("--------------")
281
+ # print("Buffer: ")
282
+ # print(" ")
283
+ # print(buffer)
284
+ # print(" ")
285
+ # print("--------------")
286
+
287
+ with torch.no_grad():
288
+ generated_ids = model.generate(input_ids['input_ids'], **generate_kwargs)
289
+ generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=True)
290
 
291
+ #generated_text = buffer
292
 
293
 
294
  return PlainTextResponse(generated_text)