-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
list of open-source publicly-available llms for code #49
Comments
diff-codegen-350m all fine-tuned from Salesforce’s CodeGen code synthesis models |
GPT-J / GPT-J-6B demo at https://6b.eleuther.ai/ |
to merge: https://github.com/eugeneyan/open-llms (section "Open LLMs for code") |
codegen2 (also supports infilling) |
https://github.com/bigcode-project/starcoder |
code-llama by Meta Code Llama: Open Foundation Models for Code |
The Mistral models https://mistral.ai/ @martinezmatias says they are good. Mistral 7B |
CodeFuse-13B: A Pretrained Multi-lingual Code Large Language Model |
QwenCODE-QWEN and CODE-QWEN-CHAT 通义千问 (Alibaba) QWEN TECHNICAL REPORT Qwen2. 5-Coder Technical Report Nov 2024:
updated @andre15silva |
DeepSeek Coder: Let the Code Write Itself |
Magicoder |
CodeShell Technical Report CodeShell-Base, a seven billion-parameter foundation model with 8K context length, showcasing exceptional proficiency in code comprehension, which outperforms CodeLlama in Humaneval after training on just 500 billion tokens (5 epochs). |
Mixtral, @FredBonux is able to use it over groq |
mistralai/Codestral-22B-v0.1 |
DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence |
aiXcoder-7B: A Lightweight and Effective Large Language Model for Code Completion |
Tencent 's Hunyuan (huggingface, paper) |
OpenCoder: The Open Cookbook for Top-Tier Code Large Language Models |
The text was updated successfully, but these errors were encountered: