title | emoji | colorFrom | colorTo | sdk | sdk_version | app_file | pinned | license | short_description |
---|---|---|---|---|---|---|---|---|---|
Small Model Chatbot |
😻 |
indigo |
green |
gradio |
5.31.0 |
app.py |
false |
mit |
Some small models chatbot |
=======
A lightweight, multi-model chat application featuring several small language models optimized for different tasks. Built with Gradio for an intuitive web interface and designed for local deployment.
- Multiple Model Support: Choose from 4 specialized small language models
- Lazy Loading: Models are loaded only when selected, optimizing memory usage
- Real-time Chat Interface: Smooth conversational experience with Gradio
- Lightweight: All models are under 200M parameters for fast inference
- Local Deployment: Run entirely on your local machine
- Purpose: General conversation and instruction following
- Architecture: HuggingFace SmolLM2-135M-Instruct
- Best For: General Q&A, creative writing, coding help
- Language: English
- Purpose: Ultra-lightweight instruction following
- Architecture: Mistral-based with chat template support
- Best For: Quick responses, simple tasks, resource-constrained environments
- Language: English
- Purpose: English to Chinese translation
- Architecture: LLaMA-based translation model
- Best For: Translating English text to Chinese
- Language: English → Chinese
- Purpose: Enhanced English to Chinese translation
- Architecture: LLaMA-based with improved accuracy
- Best For: High-quality English to Chinese translation
- Language: English → Chinese
- Python 3.8 or higher
- 4GB+ RAM recommended
- Internet connection for initial model downloads
-
Run the application
uv run app.py
-
Open your browser
- Navigate to
http://localhost:7860
- Select a model and start chatting!
- Navigate to
- Use SmolLM2 or NanoLM-25M for general chat, Q&A, and assistance
- Use NanoTranslator-S for quick English→Chinese translations
- Use NanoTranslator-XL for higher quality English→Chinese translations
- NanoLM-25M (25M params) for ultra-lightweight deployment
- NanoTranslator-S (9M params) for minimal translation needs
Model | Parameters | Use Case | Memory Usage | Speed |
---|---|---|---|---|
SmolLM2 | 135M | General Chat | ~500MB | Fast |
NanoLM-25M | 25M | Lightweight Chat | ~100MB | Very Fast |
NanoTranslator-S | 9M | Quick Translation | ~50MB | Very Fast |
NanoTranslator-XL | 78M | Quality Translation | ~300MB | Fast |
- SmolLM2:
HuggingFaceTB/SmolLM2-135M-Instruct
- NanoLM-25M:
Mxode/NanoLM-25M-Instruct-v1.1
- NanoTranslator-S:
Mxode/NanoTranslator-S
- NanoTranslator-XL:
Mxode/NanoTranslator-XL
This project is licensed under the MIT License - see the LICENSE file for details.
- HuggingFace for the Transformers library and model hosting
- Mxode for the Nano series models
- Gradio for the amazing web interface framework