Tagged "tutorial"
- A Journey to a Reliable and Enjoyable Locally Hosted Voice Assistant
- How to Build a Self-Hosted AI Server with LM Studio: Step-by-Step Guide
- Setting Up a Private AI Brain on Windows: Complete Guide to Local LLM Deployment
- Automating Read-It-Later Workflows with Local LLMs for Overnight Summarization
- Self-Hosted AI Code Review with Local LLMs: Secure Automation Guide
- Pydantic-Deep: Production Deep Agents for Pydantic AI
- Local AI Coding Assistant: Free Cursor Alternative with VS Code, Ollama & Continue
- Build a $1,500 AI Server with DeepSeek-R1 on RTX 4090
- Community Converges on Optimal KV Cache Quantization Strategies for Qwen 3.5 Models
- You're Using Your Local LLM Wrong If You're Prompting It Like a Cloud LLM
- Run LLMs Locally with Llama.cpp
- How I Used Lima for an AI Coding Agent Sandbox
- Practical Fix for Qwen 3.5 Overthinking in llama.cpp
- Show HN: Voice-tracked teleprompter using on-device ASR in the browser
- Qwen3.5-397B Achieves 282 tok/s on 4x RTX PRO 6000 Blackwell Through Custom CUTLASS Kernel
- I made Karpathy's Autoresearch work on CPU
- Local LLMs on Apple Silicon Mac 2026: M1 M2 M3 Guide
- How to Run Local LLMs in 2026: The Complete Developer's Guide
- How to Install OpenClaw with Ollama (Step-by-Step Tutorial)
- Quantization Explained: Q4_K_M vs AWQ vs FP16 for Local LLMs
- The $1,500 Local AI Setup: DeepSeek-R1 on Consumer Hardware
- Local AI Coding Assistant: Complete VS Code + Ollama + Continue Setup
- 8 Local LLM Settings Most People Never Touch That Fixed My Worst AI Problems
- How to Run Your Own Local LLM — 2026 Edition
- Llama.cpp Prompt Processing Optimization: Ubatch Size Configuration Guide
- Self-Hosted Paperless-ngx With Optional Local AI Integration
- Turning Your Linux Terminal into a Local AI Assistant
- Jse v2.0 AI Output Specification
- How to Run High-Performance LLMs Locally on the Arduino UNO Q
- 5 Useful Docker Containers for Agentic Developers
- Accuracy vs. Speed in Local LLMs: Finding Your Sweet Spot
- 5 Useful Docker Containers for Agentic Developers
- Running LLMs on Raspberry Pi and Edge Devices: A Practical Guide
- Every agent framework has the same bug – prompt decay. Here's a fix
- Building a Privacy-Preserving RAG System in the Browser
- Ollama for JavaScript Developers: Building AI Apps Without API Keys
- The Complete Developer's Guide to Running LLMs Locally: From Ollama to Production
- Qwen3.5-27B Identified as Sweet Spot for Mid-Range Local Deployment
- The Complete Stack for Local Autonomous Agents: From GGML to Orchestration
- Breaking the Speed Limit: Strategies for 17k Tokens/Sec Local Inference
- I Thought I Needed a GPU to Run AI Until I Learned About These Models
- Ollama Production Deployment: Docker-Compose Setup Guide
- AI Integration in Sublime Text: Practical Local LLM Editor Enhancement
- Running Local LLMs and VLMs on Arduino UNO Q with yzma
- Local-First RAG: Vector Search in SQLite with Hamming Distance
- Ask HN: How Do You Debug Multi-Step AI Workflows When the Output Is Wrong?
- Self-Hosted AI: A Complete Roadmap for Beginners
- Qwen3-Next 80B MoE Achieves 39 Tokens/Second on RTX 5070/5060 Ti Dual-GPU Setup
- InitRunner: YAML-Based AI Agent Framework with RAG and Memory
- Switching From Ollama and LM Studio to llama.cpp: Performance Benefits
- Optimal llama.cpp Settings Found for Qwen3 Coder Next Loop Issues
- Running Your Own AI Assistant for €19/Month: Complete Self-Hosting Guide
- OpenClaw with vLLM Running for Free on AMD Developer Cloud
- 5 Practical Ways to Use Local LLMs with MCP Tools