diff --git a/pyproject.toml b/pyproject.toml index fda987b..5ff3c6c 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta" [project] name = "llama-cpp-agent" -version = "0.2.9" +version = "0.2.10" description = "A framework for building LLM based AI agents with llama.cpp." readme = "ReadMe.md" diff --git a/src/llama_cpp_agent/llm_agent.py b/src/llama_cpp_agent/llm_agent.py index 71228ed..4682686 100644 --- a/src/llama_cpp_agent/llm_agent.py +++ b/src/llama_cpp_agent/llm_agent.py @@ -382,9 +382,9 @@ def get_response_role_and_completion( ): if len(chat_history.get_chat_messages()) == 0: if system_prompt: - chat_history.add_message({"role": "system", "content": system_prompt}) + chat_history.add_message({"role": Roles.system, "content": system_prompt}) else: - chat_history.add_message({"role": "system", "content": self.system_prompt}) + chat_history.add_message({"role": Roles.system, "content": self.system_prompt}) if message is not None and add_message_to_chat_history: chat_history.add_message(