Tagged "apple-silicon"
- Ditching Paid AI Services: Building Self-Hosted LLM Solutions as ChatGPT, Claude, and Gemini Alternatives
- Multi-Token Prediction support coming to MLX-LM for Qwen 3.5
- Apple M5 Max 128GB real-world performance benchmarks for local inference
- DeepSeek R1 RTX 4090 vs Apple M3 Max: Benchmark & Performance Guide
- NVIDIA Nemotron 3 Nano 4B Enables On-Device Inference Directly in Web Browsers via WebGPU
- Dictare – Open-source Voice Layer for AI Coding Agents (100% Local)
- Startup Transforms Mac Mini Into Full-Powered AI Inference System With External GPU
- Local LLMs on Apple Silicon Mac 2026: M1 M2 M3 Guide
- Apple M5 Max 128GB Benchmark Results for Local LLM Inference
- Experiment: 0.8B Model Self-Improvement on MacBook Air Yields Surprising Results
- M5 Max and M5 Ultra Chipsets Demonstrate Significant Bandwidth Improvements for Local LLM Inference
- Apple Launches MacBook Neo with A18 Pro Chip for Affordable Local AI Inference
- Real-World Qwen 3.5 9B Agent Performance on M1 Pro Validates Edge Deployment
- MediaTek Advances Omni Model for Efficient Smartphone Inference
- Apple Unveils MacBook Pro with M5 Pro and M5 Max Featuring On-Device AI
- Apple Unveils MacBook Pro With M5 Pro and M5 Max for On-Device AI
- Apple M5 Pro and M5 Max: 4× Faster LLM Processing
- AMD Launches Copilot+ Desktop Chips to Compete in On-Device AI Market
- VibeWhisper – macOS Voice-to-Text with 100% Local Processing Option
- Apple M4 iPad Air Targets AI Users with Double M1 Speed Performance
- Alibaba's Qwen 3.5 Small Model Runs Directly on iPhone 17
- Running Local AI Models on Mac Studio 128GB: 4B, 20B & 120B Tested
- Apple Neural Engine Reverse-Engineered for Local Model Training on Mac Mini M4
- Show HN: Caret – Tab to Complete at Any App on Your Mac
- Researchers Develop Persistent Memory System for Local LLMs—No RAG Required
- Apple: Python bindings for access to the on-device Apple Intelligence model
- Apple Accelerates U.S. Manufacturing with Mac Mini Production
- Qwen3-Code-Next Proves Practical for Local Development: Real-World Coding Tasks on Mac Studio
- Nvidia Could Launch Its First Laptops With Its Own Processors
- AI-Powered Reverse-Engineering of Rosetta 2 for Linux
- Apple Researchers Develop On-Device AI Agent That Interacts With Apps for You
- PaddleOCR-VL Now Integrated into llama.cpp for Multilingual OCR
- Complete Offline AI System: Voice Control and Smart Home via Local LLM and Radio Without Internet
- Kitten TTS V0.8 Released: State-of-the-Art Super-Tiny Text-to-Speech Model Under 25MB
- GPT4All Replaces Ollama On Mac After Quick Trial
- Meet Sarvam Edge: India's AI Model That Runs on Phones and Laptops With No Internet
- Sourdine: Open-Source macOS App for 100% Local AI Transcription
- MiniMax Releases M2.5 Model with SOTA Coding and Agent Capabilities
- MiniMax-M2.5 230B MoE Model Released with GGUF Support for Local Deployment