diff --git a/docs/examples_notebooks/drift_search.ipynb b/docs/examples_notebooks/drift_search.ipynb index 8d53c7d9cc..4e24e005fd 100644 --- a/docs/examples_notebooks/drift_search.ipynb +++ b/docs/examples_notebooks/drift_search.ipynb @@ -19,10 +19,10 @@ "import os\n", "\n", "import pandas as pd\n", - "from graphrag.config.enums import ModelType\n", + "from graphrag_llm.completion import create_completion\n", + "from graphrag_llm.config import ModelConfig, RetryConfig\n", + "from graphrag_llm.embedding import create_embedding\n", "from graphrag.config.models.drift_search_config import DRIFTSearchConfig\n", - "from graphrag.config.models.language_model_config import LanguageModelConfig\n", - "from graphrag.language_model.manager import ModelManager\n", "from graphrag.query.indexer_adapters import (\n", " read_indexer_entities,\n", " read_indexer_relationships,\n", @@ -34,7 +34,6 @@ " DRIFTSearchContextBuilder,\n", ")\n", "from graphrag.query.structured_search.drift_search.search import DRIFTSearch\n", - "from graphrag.tokenizer.get_tokenizer import get_tokenizer\n", "from graphrag_vectors.lancedb import LanceDBVectorStore\n", "\n", "INPUT_DIR = \"./inputs/operation dulce\"\n", @@ -88,7 +87,7 @@ "\n", "report_df = pd.read_parquet(f\"{INPUT_DIR}/{COMMUNITY_REPORT_TABLE}.parquet\")\n", "reports = read_indexer_reports(report_df, community_df, COMMUNITY_LEVEL)\n", - "read_indexer_report_embeddings(reports, full_content_embedding_store)" + "read_indexer_report_embeddings(reports, full_content_embedding_store)\n" ] }, { @@ -99,34 +98,24 @@ "source": [ "api_key = os.environ[\"GRAPHRAG_API_KEY\"]\n", "\n", - "chat_config = LanguageModelConfig(\n", - " api_key=api_key,\n", - " type=ModelType.Chat,\n", + "chat_config = ModelConfig(\n", " model_provider=\"openai\",\n", " model=\"gpt-4.1\",\n", - " max_retries=20,\n", - ")\n", - "chat_model = ModelManager().get_or_create_chat_model(\n", - " name=\"local_search\",\n", - " model_type=ModelType.Chat,\n", - " config=chat_config,\n", + " api_key=api_key,\n", + " retry=RetryConfig(max_retries=20),\n", ")\n", + "chat_model = create_completion(chat_config)\n", "\n", - "tokenizer = get_tokenizer(chat_config)\n", + "tokenizer = chat_model.tokenizer\n", "\n", - "embedding_config = LanguageModelConfig(\n", - " api_key=api_key,\n", - " type=ModelType.Embedding,\n", + "embedding_config = ModelConfig(\n", " model_provider=\"openai\",\n", " model=\"text-embedding-3-large\",\n", - " max_retries=20,\n", + " api_key=api_key,\n", + " retry=RetryConfig(max_retries=20),\n", ")\n", "\n", - "text_embedder = ModelManager().get_or_create_embedding_model(\n", - " name=\"local_search_embedding\",\n", - " model_type=ModelType.Embedding,\n", - " config=embedding_config,\n", - ")" + "text_embedder = create_embedding(embedding_config)\n" ] }, { diff --git a/docs/examples_notebooks/global_search.ipynb b/docs/examples_notebooks/global_search.ipynb index 605f704bd2..b0218ac1cb 100644 --- a/docs/examples_notebooks/global_search.ipynb +++ b/docs/examples_notebooks/global_search.ipynb @@ -19,9 +19,8 @@ "import os\n", "\n", "import pandas as pd\n", - "from graphrag.config.enums import ModelType\n", - "from graphrag.config.models.language_model_config import LanguageModelConfig\n", - "from graphrag.language_model.manager import ModelManager\n", + "from graphrag_llm.completion import create_completion\n", + "from graphrag_llm.config import ModelConfig, RetryConfig\n", "from graphrag.query.indexer_adapters import (\n", " read_indexer_communities,\n", " read_indexer_entities,\n", @@ -30,8 +29,7 @@ "from graphrag.query.structured_search.global_search.community_context import (\n", " GlobalCommunityContext,\n", ")\n", - "from graphrag.query.structured_search.global_search.search import GlobalSearch\n", - "from graphrag.tokenizer.get_tokenizer import get_tokenizer" + "from graphrag.query.structured_search.global_search.search import GlobalSearch\n" ] }, { @@ -58,20 +56,15 @@ "source": [ "api_key = os.environ[\"GRAPHRAG_API_KEY\"]\n", "\n", - "config = LanguageModelConfig(\n", - " api_key=api_key,\n", - " type=ModelType.Chat,\n", + "config = ModelConfig(\n", " model_provider=\"openai\",\n", " model=\"gpt-4.1\",\n", - " max_retries=20,\n", - ")\n", - "model = ModelManager().get_or_create_chat_model(\n", - " name=\"global_search\",\n", - " model_type=ModelType.Chat,\n", - " config=config,\n", + " api_key=api_key,\n", + " retry=RetryConfig(max_retries=20),\n", ")\n", + "model = create_completion(config)\n", "\n", - "tokenizer = get_tokenizer(config)" + "tokenizer = model.tokenizer\n" ] }, { diff --git a/docs/examples_notebooks/global_search_with_dynamic_community_selection.ipynb b/docs/examples_notebooks/global_search_with_dynamic_community_selection.ipynb index 6b3763d73b..5673035ab8 100644 --- a/docs/examples_notebooks/global_search_with_dynamic_community_selection.ipynb +++ b/docs/examples_notebooks/global_search_with_dynamic_community_selection.ipynb @@ -19,9 +19,8 @@ "import os\n", "\n", "import pandas as pd\n", - "from graphrag.config.enums import ModelType\n", - "from graphrag.config.models.language_model_config import LanguageModelConfig\n", - "from graphrag.language_model.manager import ModelManager\n", + "from graphrag_llm.completion import create_completion\n", + "from graphrag_llm.config import ModelConfig, RetryConfig\n", "from graphrag.query.indexer_adapters import (\n", " read_indexer_communities,\n", " read_indexer_entities,\n", @@ -30,7 +29,7 @@ "from graphrag.query.structured_search.global_search.community_context import (\n", " GlobalCommunityContext,\n", ")\n", - "from graphrag.query.structured_search.global_search.search import GlobalSearch" + "from graphrag.query.structured_search.global_search.search import GlobalSearch\n" ] }, { @@ -55,24 +54,17 @@ "metadata": {}, "outputs": [], "source": [ - "from graphrag.tokenizer.get_tokenizer import get_tokenizer\n", - "\n", "api_key = os.environ[\"GRAPHRAG_API_KEY\"]\n", "\n", - "config = LanguageModelConfig(\n", - " api_key=api_key,\n", - " type=ModelType.Chat,\n", + "config = ModelConfig(\n", " model_provider=\"openai\",\n", " model=\"gpt-4.1\",\n", - " max_retries=20,\n", - ")\n", - "model = ModelManager().get_or_create_chat_model(\n", - " name=\"global_search\",\n", - " model_type=ModelType.Chat,\n", - " config=config,\n", + " api_key=api_key,\n", + " retry=RetryConfig(max_retries=20),\n", ")\n", + "model = create_completion(config)\n", "\n", - "tokenizer = get_tokenizer(config)" + "tokenizer = model.tokenizer\n" ] }, { diff --git a/docs/examples_notebooks/local_search.ipynb b/docs/examples_notebooks/local_search.ipynb index f7f0c5a54b..b0c4bedd61 100644 --- a/docs/examples_notebooks/local_search.ipynb +++ b/docs/examples_notebooks/local_search.ipynb @@ -190,41 +190,30 @@ "metadata": {}, "outputs": [], "source": [ - "from graphrag.config.enums import ModelType\n", - "from graphrag.config.models.language_model_config import LanguageModelConfig\n", - "from graphrag.language_model.manager import ModelManager\n", - "from graphrag.tokenizer.get_tokenizer import get_tokenizer\n", + "from graphrag_llm.completion import create_completion\n", + "from graphrag_llm.config import ModelConfig, RetryConfig\n", + "from graphrag_llm.embedding import create_embedding\n", "\n", "api_key = os.environ[\"GRAPHRAG_API_KEY\"]\n", "\n", - "chat_config = LanguageModelConfig(\n", - " api_key=api_key,\n", - " type=ModelType.Chat,\n", + "chat_config = ModelConfig(\n", " model_provider=\"openai\",\n", " model=\"gpt-4.1\",\n", - " max_retries=20,\n", - ")\n", - "chat_model = ModelManager().get_or_create_chat_model(\n", - " name=\"local_search\",\n", - " model_type=ModelType.Chat,\n", - " config=chat_config,\n", + " api_key=api_key,\n", + " retry=RetryConfig(max_retries=20),\n", ")\n", + "chat_model = create_completion(chat_config)\n", "\n", - "embedding_config = LanguageModelConfig(\n", - " api_key=api_key,\n", - " type=ModelType.Embedding,\n", + "embedding_config = ModelConfig(\n", " model_provider=\"openai\",\n", " model=\"text-embedding-3-small\",\n", - " max_retries=20,\n", + " api_key=api_key,\n", + " retry=RetryConfig(max_retries=20),\n", ")\n", "\n", - "text_embedder = ModelManager().get_or_create_embedding_model(\n", - " name=\"local_search_embedding\",\n", - " model_type=ModelType.Embedding,\n", - " config=embedding_config,\n", - ")\n", + "text_embedder = create_embedding(embedding_config)\n", "\n", - "tokenizer = get_tokenizer(chat_config)" + "tokenizer = chat_model.tokenizer\n" ] }, {