Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -267,28 +267,28 @@ def simple_chat(message: dict, temperature: float = 0.8, max_length: int = 4096,
|
|
267 |
eos_token_id=[151329, 151336, 151338],
|
268 |
)
|
269 |
|
270 |
-
gen_kwargs = {**input_ids, **generate_kwargs}
|
271 |
-
|
272 |
-
with torch.no_grad():
|
273 |
-
thread = Thread(target=model.generate, kwargs=gen_kwargs)
|
274 |
-
thread.start()
|
275 |
-
buffer = ""
|
276 |
-
for new_text in streamer:
|
277 |
-
buffer += new_text
|
278 |
-
yield buffer
|
279 |
-
|
280 |
-
print("--------------")
|
281 |
-
print("Buffer: ")
|
282 |
-
print(" ")
|
283 |
-
print(buffer)
|
284 |
-
print(" ")
|
285 |
-
print("--------------")
|
286 |
|
287 |
# with torch.no_grad():
|
288 |
-
#
|
289 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
290 |
|
291 |
-
generated_text = buffer
|
292 |
|
293 |
|
294 |
return PlainTextResponse(generated_text)
|
|
|
267 |
eos_token_id=[151329, 151336, 151338],
|
268 |
)
|
269 |
|
270 |
+
# gen_kwargs = {**input_ids, **generate_kwargs}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
271 |
|
272 |
# with torch.no_grad():
|
273 |
+
# thread = Thread(target=model.generate, kwargs=gen_kwargs)
|
274 |
+
# thread.start()
|
275 |
+
# buffer = ""
|
276 |
+
# for new_text in streamer:
|
277 |
+
# buffer += new_text
|
278 |
+
# yield buffer
|
279 |
+
|
280 |
+
# print("--------------")
|
281 |
+
# print("Buffer: ")
|
282 |
+
# print(" ")
|
283 |
+
# print(buffer)
|
284 |
+
# print(" ")
|
285 |
+
# print("--------------")
|
286 |
+
|
287 |
+
with torch.no_grad():
|
288 |
+
generated_ids = model.generate(input_ids['input_ids'], **generate_kwargs)
|
289 |
+
generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=True)
|
290 |
|
291 |
+
#generated_text = buffer
|
292 |
|
293 |
|
294 |
return PlainTextResponse(generated_text)
|