Inferless
Popular repositories Loading
-
triton-co-pilot
triton-co-pilot PublicGenerate Glue Code in seconds to simplify your Nvidia Triton Inference Server Deployments
-
qwq-32b-preview
qwq-32b-preview Public templateA 32B experimental reasoning model for advanced text generation and robust instruction following. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
-
whisper-large-v3
whisper-large-v3 Public templateState‑of‑the‑art speech recognition model for English, delivering transcription accuracy across diverse audio scenarios. <metadata> gpu: T4 | collections: ["CTranslate2"] </metadata>
-
deepseek-r1-distill-qwen-32b
deepseek-r1-distill-qwen-32b Public templateA distilled DeepSeek-R1 variant built on Qwen2.5-32B, fine-tuned with curated data for enhanced performance and efficiency. <metadata> gpu: A100 | collections: ["vLLM"] </metadata>
Repositories
- kyutai-tts-1.6b Public
inferless/kyutai-tts-1.6b’s past year of commit activity - llama-3.1-8b-instruct-gguf Public template
An 8B-parameter, instruction-tuned variant of Meta's Llama-3.1 model, optimized in GGUF format for efficient inference. <metadata> gpu: A100 | collections: ["lama.cpp"] </metadata>
inferless/llama-3.1-8b-instruct-gguf’s past year of commit activity - stable-diffusion-3-5-large-turbo Public template
A fast, optimized diffusion model that generates high-quality images from text prompts, ideal for creative visual content. <metadata> gpu: A100 | collections: ["Diffusers"] </metadata>
inferless/stable-diffusion-3-5-large-turbo’s past year of commit activity - jina-embeddings-v4 Public template
A 3.8B multimodal-multilingual embedding that unifies text and image understanding in a single late-interaction space, delivers both dense and multi-vector outputs. <metadata> gpu: A10 | collections: ["HF_Transformers"] </metadata>
inferless/jina-embeddings-v4’s past year of commit activity - flux-1-kontext-dev Public template
12B model from Black Forest Labs that allows in‑context image editing with character and style consistency; supporting iterative, instruction-guided edits. <metadata> gpu: A100 | collections: ["HF_Transformers"] </metadata>
inferless/flux-1-kontext-dev’s past year of commit activity - gemma-3n-e4b-it Public template
8B variant of the lightweight Gemma 3n series that operates with a 4B‑parameter memory footprint, enabling full multimodal inference (text, image, audio, video) on resource‑constrained hardware. <metadata> gpu: A100 | collections: ["HF_Transformers"] </metadata>
inferless/gemma-3n-e4b-it’s past year of commit activity - qwen3-embedding-0.6b Public template
600M parameter, 100 language embedding model that turns up to 32k token inputs into instruction-aware vectors. <metadata> gpu: A10 | collections: ["HF_Transformers"] </metadata>
inferless/qwen3-embedding-0.6b’s past year of commit activity - devstral-small Public template
An agentic LLM for software engineering tasks, excels at using tools to explore codebases, editing multiple files and power software engineering agents. <metadata> gpu: A100 | collections: ["HF_Transformers"] </metadata>
inferless/devstral-small’s past year of commit activity - deepseek-r1-qwen3-8b Public template
A distilled 8B parameter reasoning powerhouse, leveraging deep chain‑of‑thought from the DeepSeek R1‑0528—delivering SOTA open‑source performance. <metadata> gpu: A100 | collections: ["HF Transformers"] </metadata>
inferless/deepseek-r1-qwen3-8b’s past year of commit activity - nanonets-ocr-s Public template
Nanonets-OCR-s that turns images or PDFs into structured Markdown capturing tables, LaTeX, captions and tags—for fast, powerful, human-readable OCR. <metadata> gpu: A10 | collections: ["HF_Transformers"] </metadata>
inferless/nanonets-ocr-s’s past year of commit activity
Top languages
Loading…
Most used topics
Loading…