Tagged "showcase"
-
Qwen3.5 Thinking Mode Can Be Disabled for Production Inference Optimization
-
Qwen3.5-27B Identified as Sweet Spot for Mid-Range Local Deployment
-
Show HN: Pluckr – LLM-Powered HTML Scraper That Caches Selectors and Auto-Heals
-
Show HN: MCP-Enabled File Storage for AI Agents, Auth via Ethereum Wallet
-
Show HN: 100% LLM Accuracy–No Fine-Tuning, JSON Only
-
Show HN: A Human-Curated, CLI-Driven Context Layer for AI Agents
-
Show HN: A Ground Up TLS 1.3 Client Written in C
-
Show HN: Dypai – Build Backends from Your IDE Using AI and MCP
-
Comparing Manual vs. AI Requirements Gathering: 2 Sentences vs. 127-Point Spec
-
Show HN: Agora – AI API Pricing Oracle with X402 Micropayments
-
Wave Field LLM Achieves O(n log n) Scaling: 825M Model Trained to 1B Parameters in 13 Hours
-
Qwen3's Voice Embeddings Enable Local Voice Cloning and Mathematical Voice Manipulation
-
Qwen3 Demonstrates Advanced Voice Cloning via Embeddings
-
Qwen3-Code-Next Proves Practical for Local Development: Real-World Coding Tasks on Mac Studio
-
Custom Portable Workstation Optimized for Local AI Inference Builds
-
Open-Source Framework Achieves Gemini 3 Deep Think Level Performance Through Local Model Scaffolding
-
nanollama: Open-Source Framework for Training Llama 3 from Scratch with One-Command GGUF Export
-
Massu: Governance Layer for AI Coding Assistants with 51 MCP Tools
-
Local GPT-OSS 20B Model Demonstrates Practical Agentic Capabilities
-
A Tool to Tell You What LLMs Can Run on Your Machine
-
GPT-OSS 20B Demonstrates Practical Agentic Capabilities Running Fully Locally
-
Gix: Go CLI for AI-Generated Commit Messages
-
FORTHought: Self-Hosted AI Stack for Physics Labs Built on OpenWebUI
-
Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
-
Show HN: The Only CLI Your AI Agent Will Need
-
Show HN: Tickr – AI Project Manager That Lives Inside Slack (Replaces Jira)
-
O-TITANS: Orthogonal LoRA Framework for Gemma 3 with Google TITANS Memory Architecture
-
At India AI Impact Summit, Intel Showcases AI PCs and Cost-Efficient Frugal AI
-
Show HN: Horizon – My AI-Powered Personal News Aggregator and Summarizer
-
CPU-Trained Language Model Outperforms GPU Baseline After 40 Hours
-
Vellium v0.3.5: Major Writing Mode Overhaul and Native KoboldCpp Support
-
Search and Analyze Documents from the DOJ Epstein Files Release with Local LLM
-
Qwen3 Coder Next Remains Effective at Aggressive Quantization Levels
-
At India AI Impact Summit, Intel Showcases Its AI PCs and Cost-Efficient Frugal AI
-
Apple Researchers Develop On-Device AI Agent That Interacts With Apps for You
-
24 Simultaneous Claude Code Agents on Local Hardware
-
Sarvam Brings AI to Feature Phones, Cars, and Smart Glasses
-
Running Local LLMs and VLMs on Arduino UNO Q with yzma
-
Enhanced Quantization Visualization Methods for Understanding LLM Compression Trade-offs
-
Complete Offline AI System: Voice Control and Smart Home via Local LLM and Radio Without Internet
-
Local Vision-Language Models for Document OCR and PII Detection in Privacy-Critical Workflows
-
Kitten TTS V0.8 Released: State-of-the-Art Super-Tiny Text-to-Speech Model Under 25MB
-
Clipthesis: Free Local App for Video Tagging and Search Across Drives
-
Aegis.rs: Open Source Rust-Based LLM Security Proxy Released
-
Show HN: Shiro.computer Static Page, Unix/NPM Shimmed to Host Claude Code
-
Sarvam AI Launches Edge Model to Challenge Major AI Players with Local-First Approach
-
OpenClaw Refactored in Go, Runs on $10 Hardware
-
Matmul-Free Language Model Trained on CPU in 1.2 Hours
-
Cloudflare Releases Agents SDK v0.5.0 with Rust-Powered Infire Engine for Edge Inference
-
Qwen3-Next 80B MoE Achieves 39 Tokens/Second on RTX 5070/5060 Ti Dual-GPU Setup
-
Show HN: PgCortex – AI enrichment per Postgres row, zero transaction blocking
-
Show HN: Inkog – Pre-flight check for AI agents (governance, loops, injection)
-
Cohere Releases Tiny Aya: Efficient 3.3B Multilingual Model for 70+ Languages
-
Sourdine: Open-Source macOS App for 100% Local AI Transcription
-
InitRunner: YAML-Based AI Agent Framework with RAG and Memory
-
GPU-Accelerated DataFrame Library for Local Inference Workloads
-
WinClaw: Windows-Native AI Assistant with Office Automation
-
First Vibecoded AI Operating System for Local Deployment
-
Running Mistral-7B on Intel NPU Achieves 12.6 Tokens/Second
-
Qwen Coder Next Shows Specialized Agent Performance
-
Microsoft MarkItDown: Document Preprocessing Tool for LLMs
-
Memio Launches AI-Powered Knowledge Hub for Android with Local Processing
-
New Header-Only C++ Benchmark Tool for Predictive Models on Raw Binary Streams
-
NAS System Achieves 18 tok/s with 80B LLM Using Only Integrated Graphics
-
Building a RAG Pipeline on 2M+ Pages: EpsteinFiles-RAG Project
-
Community Member Builds 144GB VRAM Local LLM Powerhouse