Tagged "edge-device"
-
Singapore's Foreign Minister Builds an AI "Second Brain" Using NanoClaw
-
Thinking Outside the Box: New Attack Surfaces in Sandboxed AI Agents
-
Show HN: Phonetic Formatter – Offline English Text to IPA on iPhone and iPad
-
75% of US Health Systems Are Using AI. Only 18% of That Deployment Is Governed
-
Google's Gemma 4 Could Put Powerful AI on Your Phone and Laptop
-
Blueprint: AI Hardware Design
-
Rust Open-Source Headless Browser for AI Agents and Web Scraping
-
Run a Local LLM Server on Raspberry Pi with Remote Access Capabilities
-
LLMs Consume 5.4x Less Mobile Energy Than Ad-Supported Web Search
-
Google's Gemma 4 Brings Powerful On-Device AI to Phones and Laptops
-
I Replaced My Local LLM With a Model Half Its Size and Got Better Results
-
Using a Local LLM as a Zero-Shot Classifier
-
I Built a Local AI Stack With 5 Docker Containers, and Now I'll Never Pay for ChatGPT Again
-
Building Real-World On-Device AI with LiteRT and NPU
-
AI Agent Designs a RISC-V CPU Core from Scratch
-
Anker Unveils 'Thus' Chip to Bring On-Device AI Across Product Line
-
Tesseron: New API Framework for AI Agents with Developer-Defined Configuration
-
Sarvam Edge: India's Offline AI Model Runs on Phones and Laptops Without Internet
-
Developer Turns Phone Into Local LLM Server with Vision, Voice, and Tool Calling Capabilities
-
16 Ways to Make a Small Language Model Think Bigger
-
Malicious GGUF Models Could Trigger Remote Code Execution on SGLang Servers
-
Gemma 4 Just Replaced My Whole Local LLM Stack
-
DeepX and Hyundai Motor Group Robotics LAB Partner to Develop Next-Generation Physical AI Compute Platform
-
ZeusHammer: Built an AI Agent That Thinks Locally
-
Controlling the Secondary Fan on Minisforum AI Pro HX 370
-
llama.cpp Merges Speculative Checkpointing for Major Inference Speed Boost
-
Intel Extends AI PC Reach With New Core Ultra Series 3 Launch
-
Bun v1.3.13
-
Waterloo's Live AI-Goose Tracker: Real-Time Edge Vision
-
Minisforum Launches N5 Max AI NAS with OpenClaw
-
LlaMa.cpp Robot Wars
-
Gemma 4 Just Replaced My Whole Local LLM Stack
-
Unweight: Lossless MLP Weight Compression for LLM Inference
-
115 TOPS in 0.67L: CHUWI AuBox X Packs On-Device AI Power Into a Palm-Sized Mini PC
-
Build a More Secure, Always-On Local AI Agent with OpenClaw and NVIDIA NemoClaw
-
The 'Ollama' Tool Has Numerous Problems, and Some Argue That Llama.cpp Is Better
-
Show HN: An MCP server that lets AI compose music on a hardware synth
-
Local AI Isn't Just Ollama—Here's the Ecosystem That Actually Makes It Useful
-
Community Computer: Collaborative Autoresearch on a Peer-to-Peer Network
-
Building a Voice AI Wearable in a Casio F91W with Whisper and BLE
-
Bonsai 1.7B in the Browser: A 290MB 1-bit LLM on WebGPU
-
Xiaomi 12 Pro Converted Into 24/7 Headless AI Server With Ollama and Gemma4
-
SigMap – Shrink AI Coding Context 97% with Auto-Scaling Token Budget
-
Self-Hosted LLMs Transform Personal Knowledge Management Systems
-
Running Gemma 4 on an iPhone 13 Pro
-
Ubiquiti UniFi G6 Turret 4K Camera Features On-Device AI Processing at $199 Price Point
-
Sovereign AI: Why the Next GPT Will Be Born in Our Living Rooms
-
Fine-Tuned Qwen3.5-0.8B for OCR Outperforms Previous 2B Release
-
Qwen 3.5 Small – On-Device Multimodal Models Released
-
OpenNebula 7.2 "Dark Horse" Released with Enhanced Infrastructure Support
-
oMLX Framework Implements DFlash Attention for Optimized Inference
-
Minisforum N5 MAX AI NAS Delivers 126 TOPS with 200TB Storage for Local LLM Workloads
-
Local LLM Connected to Home Assistant via MCP Now Enables Autonomous Smart Home Management
-
Show HN: SkillCompass – Open-Source Quality Evaluator for Your AI Skills
-
Defender – Local Prompt Injection Detection for AI Agents
-
Learn LLM Internals
-
Researchers Achieve 1-Bit Quantization of OLMo-3 7B Using Distillation
-
ASUS Malaysia to Bring UGen300 USB AI Accelerator in Q2 for Portable On-Device AI Inferencing
-
Unsloth Completes Comprehensive MiniMax M2.7 GGUF Quantization Suite
-
A Deep Dive into Tinygrad AI Compiler
-
MiniMax M2.7 Released: New Model Available for Local Deployment
-
MiniMax M2.7 Is Now Open Source
-
Google's Gemma 4 Brings Free Agentic AI to Your Phone With Zero Data Leaving the Device
-
Google Gemma 4 Delivers Exceptional Speed and Accuracy for Local Inference
-
DFlash Speculative Decoding Achieves 3.3x Speedup on Apple Silicon
-
The Best Local AI Model for Home Assistant Isn't Always the Biggest One
-
Qualcomm Snapdragon XR Powers Next-Generation AI Glasses with Local Inference
-
Google's Gemini Nano 4 Offers Faster, Smarter Local Inference Capabilities
-
DMax: New Parallel Decoding Paradigm for Diffusion Language Models
-
ASUS ExpertBook P1 Integrates On-Device AI for Enterprise Collaboration
-
AI PC Market Projected to Reach $235B by 2032, Driven by On-Device Computing Adoption
-
Self-Installing Skill Manager for AI Agents
-
Warp Decode vs. vLLM's Triton Kernel: Performance Crossover Analysis
-
Tether Launches QVAC SDK for Cross-Platform Local AI Development
-
Samsung Integrates On-Device AI Features into Galaxy A-Series Smartphones
-
Building Offline AI Companions on Severely Constrained Hardware (8GB RAM)
-
LLM Wiki v2: Extended Knowledge Base for LLM Practitioners
-
CarryAI's Serverless Vision-Language Models Enable On-Device Multimodal AI
-
Energy Consumption: The Final Frontier for AI and Local Inference
-
Speculative Decoding Made My Local LLM Actually Usable
-
Running a 1.7B Parameters LLM on an Apple Watch
-
Run Qwen3.5 on an Old Laptop: A Lightweight Local Agentic AI Setup Guide
-
Gemini-CLI, Llama.cpp, and Qwen3.5 Running on NVIDIA Jetson TK1
-
Gemma 4 Support Stabilized in Llama.cpp
-
Gemma 4 GGUF Models Updated with Critical Quantization Fixes
-
Google AI Edge Gallery Showcases Offline Inference with Gemma 4
-
Google's Gemma 4 Brings Powerful On-Device AI to Android and iOS
-
Quansloth Using Google's Turboquant Breaks the VRAM Wall for Local LLMs
-
PyTorch Foundation Welcomes Helion as a Foundation-Hosted Project to Standardize Open, Portable, and Accessible AI Kernel Authoring
-
MemPalace, the Highest-Scoring AI Memory System Ever Benchmarked
-
CricketBrain: Neuromorphic Signal Processor in Rust (0.175us/step, 944 bytes)
-
VLA Learns How to Act. S2S Decides Whether the Motion Is Physically Trustworthy
-
Verbatim 140W GAN: One of the First Chargers With USB PD 3.2 AVS (SPR) Support
-
TurboQuant in Llama.cpp Achieves 6X Smaller KV Cache
-
Show HN: Lightweight LLM Tracing Tool with CLI
-
Lenovo Korea Launches AI-Powered Industrial Edge Solutions
-
HunyuanOCR 1B: High-Quality OCR Now Viable on Budget Consumer Hardware
-
GPU Memory for LLM Inference (Part 1)
-
Google AI Edge Gallery Tops App Store Charts with On-Device Gemma 4
-
Real-time Multimodal AI on Apple Silicon: Gemma E2B Demo Shows Practical Edge Deployment
-
Apple Brings Enhanced On-Device AI Features to iPhone
-
Vektor – Local-First Associative Memory for AI Agents
-
Satsgate: Monetize AI Agents and APIs with Lightning L402 Protocol
-
Qualcomm Snapdragon Innovations Enable Advanced On-Device AI for Wearables
-
Microsoft Quantum Development Kit Ported to Rust: 100x Faster and Smaller
-
Google Previews Gemini Nano 4 for Android AICore with On-Device Capabilities
-
GMKtec NucBox K17 Launches with 97 TOPS AI Performance for Local Inference
-
Gemma 4 31B Achieves Third Place on FoodTruck Bench, Beating Larger Models
-
Gemma 4 26B MoE Emerges as Optimal All-Around Local Model for Consumer Hardware
-
Run AutoGEN with Ollama and LiteLLM in Simple Steps
-
Nex Life Logger: Local Activity Tracker with AI Agent Integration
-
Netflix Open-Sources VOID Model for Video Object Deletion
-
Kokoro TTS Achieves 20× Realtime Speed on CPU-Only On-Device Inference
-
GPUs vs. TPUs: Decoding the Powerhouses of AI
-
Google Launches Gemma 4 For Advanced On-Device AI
-
Gemma 4 KV Cache Memory Issues Fixed in llama.cpp
-
SkillCompass – Diagnose and Improve AI Agent Skills Across 6 Dimensions
-
Google Gemma 4 Released with GGUF Quantizations
-
Google Launches Gemma 4 Open Models for Local On-Device AI
-
Gemma 4 Makes Local AI Agents Practical
-
Gemma 4 2B Successfully Runs on Raspberry Pi 5
-
Gemma 4 on Arm: Optimized On-Device AI for Mobile and Edge Deployment
-
AMD Provides Day 0 Support for Gemma 4 on Ryzen AI Processors and GPUs
-
SmolLM2-360M Running on Samsung Galaxy Watch 4 with 74% Memory Reduction
-
Qwen 3.6-Plus Released
-
Show HN: Memsearch – Persistent, Cross-Agent, Cross-Session Memory for AI Agents
-
Lotte Innovate and DeepX Collaborate on Mass Production of Domestic AI Semiconductors
-
A Journey to a Reliable and Enjoyable Locally Hosted Voice Assistant
-
Show HN: Extra-Platforms, Python Library to Detect OS, Arch, Shell, CI, AI
-
Bonsai 1-Bit Models Deliver Exceptional Local Inference Performance
-
If Your AI Agent Ran NPM Install During the Axios Attack, You're Compromised
-
Local AI Ecosystem Extends Far Beyond Ollama
-
Claw64 – Full Agentic Loop in <4KB on Commodore 64
-
PrismML Announces 1-Bit Bonsai: First Commercially Viable 1-Bit LLMs
-
Running AI on a Raspberry Pi, Part 2: Running AI on a Pi in Under 5 minutes
-
Dell Technologies Unveils 10 AI PC Models for Business, from Ultralight Laptops to Ultracompact Desktops
-
TurboQuant: Understanding the Quantization Breakthrough
-
Google's TurboQuant Shows Memory Constraints Remain Critical for Local LLM Inference
-
OLED Emerges as the Display Standard for Energy-Efficient AI Systems
-
IBM Granite 4.0 3B Vision: Compact Enterprise-Grade Document AI
-
ESP32-S31: 320MHz 2-Core Microcontroller with 512KB SRAM and Networking
-
HP Launches Copilot+ PCs in India with On-Device AI Capabilities for Local Inference
-
GPU Passthrough to LXCs in Proxmox Simplifies Local LLM Deployment
-
CERN Embeds Tiny AI Models in Silicon Chips for Real-Time LHC Data Filtering
-
This Wearable Runs an On-Device AI With 2-Week Battery Life
-
Comparison of Two Frameworks: 40% Token Efficiency Improvement
-
Mistral AI Releases Voxtral: Open-Source TTS Model Beating ElevenLabs on Local Hardware
-
Apple Gets Full Gemini Access and Uses Distillation to Build Lightweight On-Device AI
-
See What Your AI Agents Are Doing: Multi-Agent Observability Tool
-
Samsung Galaxy A37 and A57 5G Launch with On-Device AI Capabilities in India
-
RF-DETR Nano and YOLO26 Enable On-Device Object Detection on Smartphones
-
Why Responsible AI Is the Bedrock of AI-Powered Applications
-
NVIDIA Releases GPT-OSS-Puzzle-88B, a Deployment-Optimized Model
-
Nota AI and SiMa.ai Partner on Physical AI Technology for Local Deployment
-
Meta Releases HyperAgents: Self-Improving AI
-
Show HN: Beforeyouship – Pre-Build Tool to Estimate LLM Cost
-
Operating Systems. One USB. ZFS on Root. AI-Powered. Free
-
Google's TurboQuant: The Unsexy AI Breakthrough Worth Watching
-
Apple Plans Slimmed-Down Gemini Models for Local iPhone AI Features
-
Google TurboQuant: Extreme Compression for Local LLM Deployment
-
Running an Open-Weight LLM Locally on an Apple Watch
-
New Open-Weight Models Released: GigaChat-3.1-Ultra and Lightning Variants
-
Lemonade 10.0.1 Improves Setup Process For Using AMD Ryzen AI NPUs On Linux
-
.APKs Are Just .ZIPs: Semi-Legally Hacking Software for Orphaned Hardware
-
Ultra-Large 400B-Class LLM Runs on iPhone in Test