Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for Local LLMs via Ollama #10

Open
Tracked by #17
chand1012 opened this issue Apr 16, 2024 · 0 comments
Open
Tracked by #17

Add support for Local LLMs via Ollama #10

chand1012 opened this issue Apr 16, 2024 · 0 comments

Comments

@chand1012
Copy link
Member

chand1012 commented Apr 16, 2024

In theory this should be possible with nothing more than the existing API, as Ollama supports an OpenAI-compatible API. However, there should be stricter token limits, which means smaller chunks for embeddings and a slew of other changes. Different models should be used for different use cases as well. For generating code, we should use starcoder2 or other similar model. For chat and text completions, Mistral 7b would be better. This should also be configurable by the user.

@chand1012 chand1012 reopened this Apr 16, 2024
@chand1012 chand1012 changed the title Add support for Local LLMs via Ollama and Llamafile Add support for Local LLMs via Ollama Apr 16, 2024
@chand1012 chand1012 mentioned this issue Apr 16, 2024
3 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant