Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Query regarding ms-graphrag.ipynb notebook #27

Open
AnimeshSingh0 opened this issue Jun 18, 2024 · 8 comments
Open

Query regarding ms-graphrag.ipynb notebook #27

AnimeshSingh0 opened this issue Jun 18, 2024 · 8 comments

Comments

@AnimeshSingh0
Copy link

I recently came across your blogs repository and noticed that you have implemented the query-focused Graphrag approach. However, it appears that the implementation is incomplete.

I am currently seeking a comprehensive implementation of this approach and was excited to find your work. Could you please let me know if you had the opportunity to complete it? If so, I would greatly appreciate it if you could share your implementation. If not, any insights into the reasons for leaving it unfinished would be highly valuable.

@tomasonjo
Copy link
Owner

tomasonjo commented Jun 18, 2024 via email

@AnimeshSingh0
Copy link
Author

Screenshot 2024-06-21 000520
When using Thread Pool executor to convert documents into graph documents, it gets stuck in between, I tried setting timeout value in future.result() and catching the timeout error, but it did not even raise the error, im not sure why does it get stuck in between, any help would be appreciated.

@tomasonjo
Copy link
Owner

tomasonjo commented Jun 21, 2024 via email

@AnimeshSingh0
Copy link
Author

The documents are 100-200 pages, and the I am trying to extract around 4 to 5 entity types with 10 possible relationships among them.

@tomasonjo
Copy link
Owner

tomasonjo commented Jun 22, 2024 via email

@AnimeshSingh0
Copy link
Author

Already doing that, tried with 500-4000 sized chunks, it works fine with less number of documents of smaller size but gets stuck when I put all the chunks for processing.

@tomasonjo
Copy link
Owner

tomasonjo commented Jun 22, 2024 via email

@AnimeshSingh0
Copy link
Author

Sure, 250 chunks, each 2000 sized

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants