Update app.py
Browse files
app.py
CHANGED
|
@@ -23,6 +23,7 @@ def respond(
|
|
| 23 |
from llama_cpp_agent import MessagesFormatterType
|
| 24 |
from llama_cpp_agent.providers import LlamaCppPythonProvider
|
| 25 |
from llama_cpp_agent.chat_history import BasicChatHistory
|
|
|
|
| 26 |
|
| 27 |
llm = Llama(
|
| 28 |
model_path="models/mistral-7b-instruct-v0.2.Q6_K.gguf",
|
|
@@ -46,13 +47,18 @@ def respond(
|
|
| 46 |
print(history)
|
| 47 |
|
| 48 |
for msn in history:
|
| 49 |
-
|
| 50 |
-
'role':
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
'content': msn[1]
|
| 52 |
}
|
| 53 |
-
print(
|
| 54 |
-
print(
|
| 55 |
-
messages.add_message(
|
|
|
|
| 56 |
|
| 57 |
stream = agent.get_chat_response(message, llm_sampling_settings=settings, chat_history=messages, returns_streaming_generator=True, print_output=False)
|
| 58 |
|
|
|
|
| 23 |
from llama_cpp_agent import MessagesFormatterType
|
| 24 |
from llama_cpp_agent.providers import LlamaCppPythonProvider
|
| 25 |
from llama_cpp_agent.chat_history import BasicChatHistory
|
| 26 |
+
from llama_cpp_agent.chat_history.messages import Roles
|
| 27 |
|
| 28 |
llm = Llama(
|
| 29 |
model_path="models/mistral-7b-instruct-v0.2.Q6_K.gguf",
|
|
|
|
| 47 |
print(history)
|
| 48 |
|
| 49 |
for msn in history:
|
| 50 |
+
user = {
|
| 51 |
+
'role': Roles.user,
|
| 52 |
+
'content': msn[0]
|
| 53 |
+
}
|
| 54 |
+
assistant = {
|
| 55 |
+
'role': Roles.assistant,
|
| 56 |
'content': msn[1]
|
| 57 |
}
|
| 58 |
+
print(user)
|
| 59 |
+
print(assistant)
|
| 60 |
+
messages.add_message(user)
|
| 61 |
+
messages.add_message(assistant)
|
| 62 |
|
| 63 |
stream = agent.get_chat_response(message, llm_sampling_settings=settings, chat_history=messages, returns_streaming_generator=True, print_output=False)
|
| 64 |
|