Update app.py
Browse files
app.py
CHANGED
@@ -23,6 +23,7 @@ def respond(
|
|
23 |
from llama_cpp_agent import MessagesFormatterType
|
24 |
from llama_cpp_agent.providers import LlamaCppPythonProvider
|
25 |
from llama_cpp_agent.chat_history import BasicChatHistory
|
|
|
26 |
|
27 |
llm = Llama(
|
28 |
model_path="models/mistral-7b-instruct-v0.2.Q6_K.gguf",
|
@@ -46,13 +47,18 @@ def respond(
|
|
46 |
print(history)
|
47 |
|
48 |
for msn in history:
|
49 |
-
|
50 |
-
'role':
|
|
|
|
|
|
|
|
|
51 |
'content': msn[1]
|
52 |
}
|
53 |
-
print(
|
54 |
-
print(
|
55 |
-
messages.add_message(
|
|
|
56 |
|
57 |
stream = agent.get_chat_response(message, llm_sampling_settings=settings, chat_history=messages, returns_streaming_generator=True, print_output=False)
|
58 |
|
|
|
23 |
from llama_cpp_agent import MessagesFormatterType
|
24 |
from llama_cpp_agent.providers import LlamaCppPythonProvider
|
25 |
from llama_cpp_agent.chat_history import BasicChatHistory
|
26 |
+
from llama_cpp_agent.chat_history.messages import Roles
|
27 |
|
28 |
llm = Llama(
|
29 |
model_path="models/mistral-7b-instruct-v0.2.Q6_K.gguf",
|
|
|
47 |
print(history)
|
48 |
|
49 |
for msn in history:
|
50 |
+
user = {
|
51 |
+
'role': Roles.user,
|
52 |
+
'content': msn[0]
|
53 |
+
}
|
54 |
+
assistant = {
|
55 |
+
'role': Roles.assistant,
|
56 |
'content': msn[1]
|
57 |
}
|
58 |
+
print(user)
|
59 |
+
print(assistant)
|
60 |
+
messages.add_message(user)
|
61 |
+
messages.add_message(assistant)
|
62 |
|
63 |
stream = agent.get_chat_response(message, llm_sampling_settings=settings, chat_history=messages, returns_streaming_generator=True, print_output=False)
|
64 |
|