Tagged "fine-tuning"
- Qwen3.5-27B Emerges as Sweet Spot for Single-GPU Local Deployment
- Building a Production AI Receptionist: Practical Local LLM Deployment Case Study
- Llama 8B Matches 70B Performance on Multi-Hop QA Using Structured Prompting
- Why You Should Use Both ChatGPT and Local LLMs: A Practical Hybrid Approach
- Self-Hosted AI Code Review with Local LLMs: Secure Automation Guide
- Ultra-Compact 28M Parameter Models Show Promise for Specialized Domain Tasks
- Cursor's Composer 2 Model Analysis – Fine-Tuned Variant of Kimi K2.5
- Tether's QVAC Introduces Cross-Platform Bitnet LoRA Framework for On-Device AI Training
- Unsloth Studio: Open-Source Web UI for Training and Running LLMs Locally
- On-Device AI: Tether's QVAC Fabric Enables Local Training
- Mistral Releases Small 4 Open-Source Model Under Apache 2.0
- Mistral Releases Leanstral: First Open-Source Code Agent for Lean 4 Proof Assistant
- Researcher Discovers Universal "Danger Zone" in Transformer Model Architecture at 50% Depth
- KAIST Develops World's First Hyper-Personalized On-Device AI Chip
- NVIDIA Updates Nemotron 3 122B License, Removes Deployment Restrictions
- Show HN: Generate, Clean, and Prepare LLM Training Data, All-in-One
- StepFun Releases SFT Dataset Used to Train Step 3.5 Flash for Community Fine-Tuning
- OpenClaw vs Eigent vs Claude Cowork: Comparing Open-Source AI Collaboration Platforms
- Fine-Tuned 14B Model Outperforms Claude Opus 4.6 on Ada Code Generation
- Sarvam Open-Sources 30B and 105B Reasoning Models
- Experiment: 0.8B Model Self-Improvement on MacBook Air Yields Surprising Results
- Texas Instruments Launches NPU-Powered MCUs for Low-Power Edge AI
- Sarvam Open-Sources 30B and 105B Reasoning Models
- LMF – LLM Markup Format
- Show HN: AIWatermarkDetector: Detect AI Watermarks in Text or Code
- .ispec: Runtime Specification Validation for AI System Consistency
- Fish Audio Open-Sources S2: Expressive Text-to-Speech with Natural Language Control and 100ms Latency
- Fine-Tuned Qwen SLMs (0.6–8B) Demonstrate Competitive Performance Against Frontier LLMs on Specialized Tasks
- Sarvam Open-Sources 30B and 105B Reasoning Models
- FretBench – Testing 14 LLMs on Reading Guitar Tabs Reveals Performance Gaps
- Apple Launches MacBook Neo with A18 Pro Chip for Affordable Local AI Inference
- Sarvam AI Releases 30B and 105B Open-Source Models Trained from Scratch
- Show HN: BoardMint – A PCB Review Tool That Avoids AI Hallucinations
- Incrmd: Incremental AI Coding by Editing PROJECT.md
- Qwen 3.5 vs Qwen 3 Benchmark Analysis: Generational Performance Improvements Visualized
- Jan Releases Code-Tuned 4B Model for Efficient Local Code Generation and Development Tasks
- Change Intent Records: The Missing Artifact in AI-Assisted Development
- Apple Neural Engine Reverse-Engineered for Local Model Training on Mac Mini M4
- Nummi – AI Companion with Memory and Daily Guidance
- Google Research Finds Longer Chain-of-Thought Correlates Negatively With Accuracy
- Extracting 100K Concepts from an 8B LLM
- Researchers Develop Persistent Memory System for Local LLMs—No RAG Required
- Show HN: 100% LLM Accuracy–No Fine-Tuning, JSON Only
- Anthropic Has Never Open-Sourced an LLM: Implications for Local Deployment Strategy
- Comparing Manual vs. AI Requirements Gathering: 2 Sentences vs. 127-Point Spec
- Wave Field LLM Achieves O(n log n) Scaling: 825M Model Trained to 1B Parameters in 13 Hours
- nanollama: Open-Source Framework for Training Llama 3 from Scratch with One-Command GGUF Export
- Local GPT-OSS 20B Model Demonstrates Practical Agentic Capabilities
- O-TITANS: Orthogonal LoRA Framework for Gemma 3 with Google TITANS Memory Architecture
- CPU-Trained Language Model Outperforms GPU Baseline After 40 Hours
- Matmul-Free Language Model Trained on CPU in 1.2 Hours
- Can We Leverage AI/LLMs for Self-Learning?
- Cohere Releases Tiny Aya: Efficient 3.3B Multilingual Model for 70+ Languages
- GPU-Accelerated DataFrame Library for Local Inference Workloads
- Developer Creates Custom Local AI Headshot Generator After Commercial Solutions Fail