diff --git a/README.md b/README.md index 18552738..66e64705 100644 --- a/README.md +++ b/README.md @@ -98,7 +98,7 @@ Once you have the above docker setup and running, you can create agents and init }, "llm_agent": { "agent_flow_type": "streaming", - "family": "openai", + "provider": "openai", "request_json": true, "model": "gpt-3.5-turbo-16k", "use_fallback": true diff --git a/bolna/providers.py b/bolna/providers.py index 22d1fbfd..bb82f295 100644 --- a/bolna/providers.py +++ b/bolna/providers.py @@ -39,7 +39,8 @@ 'anyscale': LiteLLM, 'custom': OpenAiLLM, 'ola': OpenAiLLM, - 'groq': LiteLLM + 'groq': LiteLLM, + 'anthropic': LiteLLM } SUPPORTED_INPUT_HANDLERS = { 'default': DefaultInputHandler, diff --git a/requirements.txt b/requirements.txt index 22c86b7e..d65221f2 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,9 +1,10 @@ aiobotocore==2.9.0 aiofiles==23.2.1 aiohttp==3.9.1 +daily-python==0.9.1 fastapi==0.108.0 fastembed==0.2.7 -litellm>=1.15 +litellm==1.40.20 numpy==1.26.1 openai==1.10.0 pydantic==2.5.3 @@ -24,5 +25,3 @@ huggingface-hub semantic-router sentence_transformers optimum[onnxruntime] -fastapi_utilities -daily-python \ No newline at end of file