Tagged "hardware"
-
Four Raspberry Pi AI Tools You Can Try This Week Beyond OpenClaw
-
Open-Source Tool Helps Determine Which Local LLMs Run on Your PC
-
Running a Private AI Brain on Windows PC as Alternative to Cloud Services
-
Llama.cpp ROCm 7 vs Vulkan Performance Benchmarks on AMD Mi50
-
Korea to Deploy Domestic AI Chips in Smart Cities as NPU Trials Scale Up
-
Powerful AI Search Engine Built on Single GeForce RTX 5090
-
Ditching Paid AI Services: Building Self-Hosted LLM Solutions as ChatGPT, Claude, and Gemini Alternatives
-
Setting Up a Private AI Brain on Windows: Complete Guide to Local LLM Deployment
-
Qualcomm and Samsung's 30-Year AI Alliance Enters a New Phase as On-Device AI Chip Race Heats Up
-
Apple M5 Max 128GB real-world performance benchmarks for local inference
-
DeepSeek R1 RTX 4090 vs Apple M3 Max: Benchmark & Performance Guide
-
Build a $1,500 AI Server with DeepSeek-R1 on RTX 4090
-
Qwen 3.5 Emerges as Top Performer for Local Deployment with Extensive Quantization Options
-
Repurpose Old GPUs as Dedicated AI Inference Accelerators
-
NVIDIA Nemotron Cascade 2 30B Delivers 120B-Class Performance in Compact Form Factor
-
NVIDIA Nemotron 3 Nano 4B Enables On-Device Inference Directly in Web Browsers via WebGPU
-
ASUS ExpertCenter PN55 Mini PC Combines AMD AI CPU and 55 TOPS NPU
-
Multiverse Computing Targets On-Device AI With Compressed Models and New API Portal
-
Dell Pro Max 16 Plus Launches With Enterprise-Grade Discrete NPU for On-Device AI
-
Snapdragon 8 Elite Gen 5 Hands the Galaxy S26 the AI Upgrade We've Been Waiting For
-
Custom GPU Multiplexer Achieves 0.3ms Model Switching on Legacy Hardware
-
Run LLMs Locally with Llama.cpp
-
A New Magnetic Material for the AI Era
-
KAIST Develops World's First Hyper-Personalized On-Device AI Chip
-
NVIDIA Updates Nemotron 3 122B License, Removes Deployment Restrictions
-
Nota Added to Three Technology and Growth ETFs in a Row – Market Recognition for AI Efficiency
-
This External GPU Enclosure Tries to Break Cloud Dependence for Local AI Inference
-
Custom AI Smart Speaker
-
AMD Declares 'AI on the PC Has Crossed an Important Line' – Agent Computers as Next Breakthrough
-
Qwen3.5-397B Achieves 282 tok/s on 4x RTX PRO 6000 Blackwell Through Custom CUTLASS Kernel
-
Nvidia's Nemotron 3 Super: Understanding the Significance for Local LLM Deployment
-
Running Qwen3.5-27B Across Multiple GPUs Over LAN Achieves Practical Speed for Local Inference
-
Startup Transforms Mac Mini Into Full-Powered AI Inference System With External GPU
-
Open-Source GreenBoost Driver Augments NVIDIA GPU VRAM With System RAM and NVMe Storage
-
I made Karpathy's Autoresearch work on CPU
-
AMD Launches Agent System Optimized for Local AI Inference With Ryzen and Radeon
-
Achieving 2000 Tokens Per Second with QWEN 3.5 27B on RTX-5090
-
Local LLMs on Apple Silicon Mac 2026: M1 M2 M3 Guide
-
Lemonade v10 Brings Linux NPU Support and Multi-Modal Capabilities
-
Linux 7.0 AMDGPU Fixing Idle Power Issue For RDNA4 GPUs After Compute Workloads
-
Intel Updates LLM-Scaler-vLLM With Support For More Qwen3/3.5 Models
-
Show HN: VmExit – An Experiment in AI-Native Computing
-
Nvidia Pushes Jetson as Edge Hub for Open AI Models
-
Comprehensive MoE Backend Benchmarks for Qwen3.5-397B: Real Numbers vs Hype
-
Apple M5 Max 128GB Benchmark Results for Local LLM Inference
-
The $1,500 Local AI Setup: DeepSeek-R1 on Consumer Hardware
-
Llama.cpp Adds True Reasoning Budget Support
-
Cutile.jl Brings Nvidia CUDA Tile-Based Programming to Julia
-
Experiment: 0.8B Model Self-Improvement on MacBook Air Yields Surprising Results
-
Texas Instruments Launches NPU-Powered MCUs for Low-Power Edge AI
-
SK Hynix Completes Qualification for LPDDR6 Memory Optimized for AI Inference
-
NVIDIA Jetson Brings Open Models to Life at the Edge
-
SK Hynix Develops 1c LPDDR6 DRAM to Boost On-Device AI Performance in Mobile Devices
-
PhotoPrism AI-Powered Photos App Brings Better Ollama Integration
-
HP OMEN MAX 16 Review: Is Local AI on a Laptop Viable in 2026?
-
Google Delivers On-Device AI Features in New Chromebook Plus Model
-
FreeBSD 14.4 Released: Implications for Local LLM Deployment
-
M5 Max and M5 Ultra Chipsets Demonstrate Significant Bandwidth Improvements for Local LLM Inference
-
Community Survey: AI Content Automation Stacks in 2026
-
Strix Halo (Ryzen AI Max+ 395) Achieves Strong Local Inference Performance with ROCm 7.2
-
When Running Ollama on Your PC for Local AI, One Thing Matters More Than Most
-
Nota AI to Showcase End-to-End On-Device AI Optimization at Embedded World 2026
-
Nemotron 9B Powers Large-Scale Local Inference: Patent Classification and Real-Time Applications
-
Gyro-Claw – Secure Execution Runtime for AI Agents
-
Engram – Open-Source Persistent Memory for AI Agents
-
Snapdragon Wear Elite Unveiled at MWC 2026, Advancing Wearable AI Inference
-
Samsung Opens Registration for Vision AI QLED and OLED Television Integration
-
HP Refreshes Lineup with AI-Focused Workstations
-
Apple Launches MacBook Neo with A18 Pro Chip for Affordable Local AI Inference
-
Windows 11 Notepad Gets On-Device AI Text Generation Without Subscription
-
Building PyTorch-Native Support for IBM Spyre Accelerator
-
The Emerging Role of SRAM-Centric Chips in AI Inference
-
Unity Showcases Manufacturing AI Workflow at Smart Factory Expo
-
MediaTek Advances Omni Model for Efficient Smartphone Inference
-
Apple Unveils MacBook Pro with M5 Pro and M5 Max Featuring On-Device AI
-
Qualcomm Snapdragon Wear Elite Brings On-Device AI to Smartwatches
-
OpenWrt 25.12.0 – Stable Release
-
On-Device AI Laptop Lineups Become Standard Across Major Manufacturers
-
Apple Unveils MacBook Pro With M5 Pro and M5 Max for On-Device AI
-
Apple M5 Pro and M5 Max: 4× Faster LLM Processing
-
AMD Launches Copilot+ Desktop Chips to Compete in On-Device AI Market
-
Qualcomm Snapdragon Wear Elite: 2B Parameter NPU for Personal AI Wearables
-
Intel Arc Pro B70 Workstation GPU Confirmed via vLLM AI Release Notes
-
Apple M4 iPad Air Targets AI Users with Double M1 Speed Performance
-
AMD Ryzen AI 400 Series Desktop Processors Launch with Integrated 60 TOPS NPU
-
Qwen 3.5 27B Achieves 100+ Tokens/s Decode on Dual RTX 3090s with 170K Context
-
Qualcomm Launches Snapdragon Wear Elite for On-Device AI on Wearables
-
Local LLM Performance Improvements: A Year of Progress Since DeepSeek R1 Moment
-
HP ZBook Ultra 14 G1a Workstation Reclaims Local AI Workflows for Professionals
-
Apple Neural Engine Reverse-Engineered for Local Model Training on Mac Mini M4
-
AMD Expands Ryzen AI 400 Series Portfolio for Consumer and Enterprise AI PC Options
-
How to Run High-Performance LLMs Locally on the Arduino UNO Q
-
Qwen 3.5-35B-A3B Emerges as Efficient Daily Driver, Replacing 120B Models
-
Huawei's SuperPoD Portfolio Creates New Option for Global Computing at MWC Barcelona 2026
-
Bare-Metal LLM Inference: UEFI Application Boots Directly Into LLM Chat
-
Qwen3.5-35B RTX 5080 Experiments Confirm KV q8_0 as Free Lunch, Q4_K_M Remains Optimal
-
Qwen3.5-35B Successfully Runs on Raspberry Pi 5 at 3+ Tokens/Second
-
Qwen 3.5-35B RTX 5080 Benchmarks Confirm KV Q8_0 as Free Lunch, Q4_K_M Remains Optimal
-
The ML.energy Leaderboard
-
Meta Reveals AI-Packed Smartwatch In 2026 – Why Wearables Shift Now
-
LLmFit: Terminal Tool for Right-Sizing LLM Models to Your Hardware
-
LLmFit: One-Command Hardware-Aware Model Selection Across 497 Models and 133 Providers
-
Krasis: Hybrid CPU/GPU MoE Runtime Achieves 3,324 Tokens/Second Prefill on RTX 5080
-
Krasis Hybrid MoE Runtime Achieves 3,324 tok/s Prefill on Single RTX 5080
-
Arduino, Qualcomm Bring On-Device AI and Robotics Learning to Indian School Systems
-
Snapdragon 8 Elite Gen 5 for Galaxy Official: 5 Key Improvements that Push the Boundaries
-
Seco Launches Edge AI System-on-Module at Embedded World 2026
-
Snapdragon 8 Elite Gen 5 Powers Galaxy S26 Series With Enhanced On-Device AI
-
Arduino and Qualcomm Bring On-Device AI Learning to Indian Schools
-
Running LLMs on Raspberry Pi and Edge Devices: A Practical Guide
-
Qwen 3.5 MoE Delivers 100K Context Window at 40+ TPS on RTX 5060 Ti
-
Qwen3.5 122B Achieves 25 tok/s on 72GB VRAM Setup
-
Researchers Develop Persistent Memory System for Local LLMs—No RAG Required
-
The Complete Developer's Guide to Running LLMs Locally: From Ollama to Production
-
New Era of On-Device AI Driven by High-Speed UFS 5.0 Storage
-
Qwen3.5-27B Identified as Sweet Spot for Mid-Range Local Deployment
-
How AI is Redefining Price and Performance in Modern Laptops
-
Kioxia Sampling UFS 5.0 Embedded Flash Memory for Next-Generation Mobile Applications
-
Apple Accelerates U.S. Manufacturing with Mac Mini Production
-
South Korea to Launch $687 Million Project to Develop On-Device AI Semiconductors
-
Qwen3-Code-Next Proves Practical for Local Development: Real-World Coding Tasks on Mac Studio
-
Custom Portable Workstation Optimized for Local AI Inference Builds
-
Nvidia Could Launch Its First Laptops With Its Own Processors
-
Yet Another Fix Coming for Older AMD GPUs on Linux – Thanks to Valve Developer
-
AI-Powered Reverse-Engineering of Rosetta 2 for Linux
-
AI Is Stress Testing Processor Architectures and RISC-V Fits the Moment
-
How Slow Local LLMs Are on My Framework 13 AMD Strix Point
-
At India AI Impact Summit, Intel Showcases AI PCs and Cost-Efficient Frugal AI
-
Google Open-Sources NPU IP, Synaptics Implements It for Hardware Acceleration
-
Asus ExpertBook B3 G2 with 50 TOPS AI Sets New Enterprise Standard
-
AI PCs Explained: 7 Critical Truths About NPUs and Privacy
-
Taalas Etches AI Models onto Transistors to Rocket Boost Inference
-
Strix Halo Performance Benchmarks: Minimax M2.5, Step 3.5 Flash, Qwen3 Coder
-
I Run Local LLMs in One of the World's Priciest Energy Markets, and I Can Barely Tell
-
At India AI Impact Summit, Intel Showcases Its AI PCs and Cost-Efficient Frugal AI
-
Google Is Exploring Ways to Use Its Financial Might to Take on Nvidia
-
Apple Researchers Develop On-Device AI Agent That Interacts With Apps for You
-
SanityBoard Adds 27 New Model Evaluations Including Qwen 3.5 Plus, GLM 5, and Gemini 3.1 Pro
-
Free ASIC-Accelerated Llama 3.1 8B Inference at 16,000 Tokens/Second
-
Complete Offline AI System: Voice Control and Smart Home via Local LLM and Radio Without Internet
-
Same INT8 Model Shows 93% to 71% Accuracy Variance Across Snapdragon Chipsets
-
Matmul-Free Language Model Trained on CPU in 1.2 Hours
-
Self-Hosted AI: A Complete Roadmap for Beginners
-
Meet Sarvam Edge: India's AI Model That Runs on Phones and Laptops With No Internet
-
Asus ExpertBook B3 G2 Laptop Features Ryzen AI 9 HX 470 CPU in 1.41kg Ultraportable Form Factor
-
LLaDA2.1 Introduces Token Editing for Massive Speed Gains in Local Inference
-
Context Management Identified as Real Bottleneck in AI-Assisted Coding
-
Ring-1T-2.5 Released with SOTA Deep Thinking Performance