Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -258,6 +258,8 @@ def simple_chat(message: dict, temperature: float = 0.8, max_length: int = 4096,
|
|
258 |
streamer=streamer
|
259 |
)
|
260 |
|
|
|
|
|
261 |
def generate_text():
|
262 |
with torch.no_grad():
|
263 |
model.generate(input_ids['input_ids'], **generate_kwargs)
|
@@ -267,10 +269,11 @@ def simple_chat(message: dict, temperature: float = 0.8, max_length: int = 4096,
|
|
267 |
thread.start()
|
268 |
|
269 |
# Collect generated text in real-time
|
270 |
-
buffer = ""
|
271 |
for new_text in streamer:
|
272 |
buffer += new_text
|
273 |
-
|
|
|
|
|
274 |
|
275 |
print(" ")
|
276 |
print("------")
|
@@ -279,9 +282,7 @@ def simple_chat(message: dict, temperature: float = 0.8, max_length: int = 4096,
|
|
279 |
print(" ")
|
280 |
print("------")
|
281 |
|
282 |
-
|
283 |
-
|
284 |
-
return PlainTextResponse(results_text)
|
285 |
except Exception as e:
|
286 |
return PlainTextResponse(f"Error: {str(e)}")
|
287 |
|
|
|
258 |
streamer=streamer
|
259 |
)
|
260 |
|
261 |
+
buffer = ""
|
262 |
+
|
263 |
def generate_text():
|
264 |
with torch.no_grad():
|
265 |
model.generate(input_ids['input_ids'], **generate_kwargs)
|
|
|
269 |
thread.start()
|
270 |
|
271 |
# Collect generated text in real-time
|
|
|
272 |
for new_text in streamer:
|
273 |
buffer += new_text
|
274 |
+
|
275 |
+
# Wait for the generation thread to finish
|
276 |
+
thread.join()
|
277 |
|
278 |
print(" ")
|
279 |
print("------")
|
|
|
282 |
print(" ")
|
283 |
print("------")
|
284 |
|
285 |
+
return PlainTextResponse(buffer)
|
|
|
|
|
286 |
except Exception as e:
|
287 |
return PlainTextResponse(f"Error: {str(e)}")
|
288 |
|