ML Researcher · Open-Source Engineer · Apple Silicon AI
Stuttgart — building the infrastructure that makes local AI actually work.
Acknowledged contributor across the core MLX stack: mlx · mlx-lm · mlx-examples · mlx-vlm
Training features added to mlx / mlx-lm: Full-weight fine-tuning · Muon optimizer · ReLU² activation · WandB reporting · Multi-optimizer support
Highlighted model architectures ported to MLX-LM
| Model | Org |
|---|---|
| Mamba v1, v2, v3 | State Space |
| MiniCPM, MiniCPM3 | OpenBMB |
| Helium | Kyutai |
| GLM, GLM4, GLM5 | Z.ai / THUKEG |
| dots.llm1 | Rednote |
| Ernie 4.5 MoE | Baidu |
| Bailing MoE, Bailing Linear (Ling) | inclusionAI |
| Granite MoE | IBM |
| LongCat | Meituan |
| Nemotron H | NVIDIA |
| Apertus | Swiss-AI |
| OLMoE, OLMo 3 | AllenAI |
| Jamba | AI21 Labs |
| ...and more | See mlx-lm commit history |
| Paper | Notes | Year |
|---|---|---|
| DynaMoE | Dynamic adaptive Mixture-of-Experts LLM architecture | 2026 |
| Gabliteration | Automated abliteration for any Transformers-compatible LLM | 2025 |
mlx-lm-lora — LoRA / QLoRA / full fine-tuning on Apple Silicon. 12+ training methods, DPO / GRPO / ORPO / PPO, Muon optimizer, WandB. The go-to fine-tuning toolkit for M-series.
mlx-vlm (main trainer maintainer) — Vision-language model training on MLX. Fully rewrote the training backend, added ORPO support.
MLX-Embeddings-LoRA — Fine-tune embedding models for retrieval and semantic tasks on Apple Silicon.
MLX-Benchmark — First CLI benchmark measuring LLM understanding of the MLX ecosystem and its APIs.
Moshi-FineTune-MLX — LoRA and full fine-tuning for Moshi speech-to-speech models on Apple Silicon.
Local NotebookLM — Fully local, PDF-grounded audio generation (up to 6 speakers). No API keys, no cloud. Companion native app included.
MLX-LM-LENS — Interpretability and abliteration tooling for MLX language models.
MLX-KAN — Kolmogorov-Arnold Networks, natively in MLX.
Gabliteration — Companion repo to arXiv:2412.06527. Remove refusal directions from any HF Transformers model.
J.O.S.I.E.-Home — Fully local real-time multimodal smart home assistant. Discrete diffusion LM, custom ChatML-style tokenizer (hardcoded vocab: rooms, devices, properties, value bins). No cloud dependency.
Josie-Linear — New Linear Dynamic Mixture-of-Experts LLM architecture.
If my work has saved you GPU bills or ended up in your pipeline — consider sponsoring. Everything here is free, maintained in my spare time.





