Update app.py
Browse files
app.py
CHANGED
|
@@ -53,13 +53,14 @@ def predict(message, history):
|
|
| 53 |
print(f'history: {history}')
|
| 54 |
for i, item in enumerate(history):
|
| 55 |
model_messages.append({"role": user_role, "content": item[0]})
|
| 56 |
-
|
| 57 |
-
model_messages.append({"role": assistant_role, "content": item[1]})
|
| 58 |
|
|
|
|
|
|
|
| 59 |
print(f'model_messages: {model_messages}')
|
| 60 |
-
|
|
|
|
| 61 |
model_inputs = tokenizer.apply_chat_template(model_messages, add_generation_prompt=True, return_tensors="pt").to(device)
|
| 62 |
-
print(f'model_final_inputs: {tokenizer.apply_chat_template(model_messages, add_generation_prompt=True, tokenize=False)}')
|
| 63 |
# model_inputs = tokenizer([messages], return_tensors="pt").to(device)
|
| 64 |
|
| 65 |
streamer = TextIteratorStreamer(tokenizer, timeout=10., skip_prompt=True, skip_special_tokens=True)
|
|
|
|
| 53 |
print(f'history: {history}')
|
| 54 |
for i, item in enumerate(history):
|
| 55 |
model_messages.append({"role": user_role, "content": item[0]})
|
| 56 |
+
model_messages.append({"role": assistant_role, "content": item[1]})
|
|
|
|
| 57 |
|
| 58 |
+
model_messages.append({"role": user_role, "content": message})
|
| 59 |
+
|
| 60 |
print(f'model_messages: {model_messages}')
|
| 61 |
+
|
| 62 |
+
print(f'model_final_inputs: {tokenizer.apply_chat_template(model_messages, add_generation_prompt=True, tokenize=False)}, flash=True')
|
| 63 |
model_inputs = tokenizer.apply_chat_template(model_messages, add_generation_prompt=True, return_tensors="pt").to(device)
|
|
|
|
| 64 |
# model_inputs = tokenizer([messages], return_tensors="pt").to(device)
|
| 65 |
|
| 66 |
streamer = TextIteratorStreamer(tokenizer, timeout=10., skip_prompt=True, skip_special_tokens=True)
|