Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -25,7 +25,7 @@ def respond(message, chat_history, system_message, max_tokens, temperature, top_
|
|
25 |
messages.append({"role": "user", "content": message})
|
26 |
|
27 |
try:
|
28 |
-
|
29 |
for chunk in client.text_generation(
|
30 |
prompt="\n".join([f"{m['role']}: {m['content']}" for m in messages]),
|
31 |
max_new_tokens=max_tokens,
|
@@ -36,9 +36,11 @@ def respond(message, chat_history, system_message, max_tokens, temperature, top_
|
|
36 |
if stop_event.is_set():
|
37 |
break
|
38 |
if chunk:
|
39 |
-
|
40 |
-
chat_history
|
41 |
-
|
|
|
|
|
42 |
except Exception as e:
|
43 |
error_message = f"오류 발생: {str(e)}"
|
44 |
chat_history.append((message, error_message))
|
|
|
25 |
messages.append({"role": "user", "content": message})
|
26 |
|
27 |
try:
|
28 |
+
full_response = ""
|
29 |
for chunk in client.text_generation(
|
30 |
prompt="\n".join([f"{m['role']}: {m['content']}" for m in messages]),
|
31 |
max_new_tokens=max_tokens,
|
|
|
36 |
if stop_event.is_set():
|
37 |
break
|
38 |
if chunk:
|
39 |
+
full_response += chunk
|
40 |
+
yield chat_history + [(message, full_response)]
|
41 |
+
|
42 |
+
chat_history.append((message, full_response))
|
43 |
+
yield chat_history
|
44 |
except Exception as e:
|
45 |
error_message = f"오류 발생: {str(e)}"
|
46 |
chat_history.append((message, error_message))
|