diff --git a/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/pyproject.toml b/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/pyproject.toml index ea28c631..460314f8 100644 --- a/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/pyproject.toml +++ b/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/pyproject.toml @@ -8,4 +8,4 @@ license = "{{cookiecutter.project_metadata.license}}" [tool.poetry.dependencies] python = "^3.11" agentops = "^0.3.12" -autogen-agentchat = "~0.2" \ No newline at end of file +autogen = "0.6.0" \ No newline at end of file diff --git a/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/src/main.py b/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/src/main.py index 4bd463c0..41c6f768 100644 --- a/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/src/main.py +++ b/agentstack/templates/autogen/{{cookiecutter.project_metadata.project_slug}}/src/main.py @@ -9,7 +9,7 @@ def main(): # Load LLM inference endpoints from an env variable or a file - # See https://microsoft.github.io/autogen/docs/FAQ#set-your-api-endpoints + # See https://docs.ag2.ai/docs/topics/llm_configuration # and OAI_CONFIG_LIST_sample. # For example, if you have created a OAI_CONFIG_LIST file in the current working directory, that file will be used. config_list = config_list_from_json(env_or_file="OAI_CONFIG_LIST")