Tagged "mlx-framework"
- I Replaced My Local LLM With a Model Half Its Size and Got Better Results
- Llama 4 Scout on MLX: The Complete Apple Silicon Guide (2026)
- DFlash Speculative Decoding Achieves 3.3x Speedup on Apple Silicon
- Kokoro TTS Achieves 20× Realtime Speed on CPU-Only On-Device Inference
- Apple Silicon Macs Run Local AI Faster with Ollama's New MLX Support
- Ollama Adopts Apple's MLX Framework for Faster Local AI on Mac
- mlx-Code: Run Claude Code Locally with MLX-LM
- Google TurboQuant: Extreme Compression for Local LLM Deployment