Tagged "training"
-
Chinese LLM Ecosystem Landscape: ByteDance Doubao, Alibaba, and Open-Source Competition
-
Building a Production AI Receptionist: Practical Local LLM Deployment Case Study
-
Local AI Coding Assistant: Free Cursor Alternative with VS Code, Ollama & Continue
-
Cursor's Composer 2 model attribution dispute highlights open-source licensing concerns
-
Your Site Content Is Powering AI. Your Bank Account Has No Idea
-
Ultra-Compact 28M Parameter Models Show Promise for Specialized Domain Tasks
-
NVIDIA Nemotron Cascade 2 30B Delivers 120B-Class Performance in Compact Form Factor
-
Cursor's Composer 2 Model Analysis – Fine-Tuned Variant of Kimi K2.5
-
Tether's QVAC Introduces Cross-Platform Bitnet LoRA Framework for On-Device AI Training
-
Unsloth Studio: Open-Source Web UI for Training and Running LLMs Locally
-
On-Device AI: Tether's QVAC Fabric Enables Local Training
-
You're Using Your Local LLM Wrong If You're Prompting It Like a Cloud LLM
-
Show HN: Generate, Clean, and Prepare LLM Training Data, All-in-One
-
StepFun Releases SFT Dataset Used to Train Step 3.5 Flash for Community Fine-Tuning
-
Fine-Tuned 14B Model Outperforms Claude Opus 4.6 on Ada Code Generation
-
Experiment: 0.8B Model Self-Improvement on MacBook Air Yields Surprising Results
-
Show HN: AIWatermarkDetector: Detect AI Watermarks in Text or Code
-
Sarvam AI Releases 30B and 105B Open-Source Models Trained from Scratch
-
Mojo: Creating a Programming Language for an AI World with Chris Lattner
-
Change Intent Records: The Missing Artifact in AI-Assisted Development
-
C7: Pipe Up-to-Date Library Docs Into Any LLM From the Terminal
-
Apple Neural Engine Reverse-Engineered for Local Model Training on Mac Mini M4
-
Google Research Finds Longer Chain-of-Thought Correlates Negatively With Accuracy
-
Arduino, Qualcomm Bring On-Device AI and Robotics Learning to Indian School Systems
-
Show HN: 100% LLM Accuracy–No Fine-Tuning, JSON Only
-
No, Local LLMs Can't Replace ChatGPT or Gemini — I Tried
-
Anthropic Has Never Open-Sourced an LLM: Implications for Local Deployment Strategy
-
Anthropic Reveals Industrial-Scale Distillation Attacks by Chinese AI Labs
-
Wave Field LLM Achieves O(n log n) Scaling: 825M Model Trained to 1B Parameters in 13 Hours
-
How Do You Know Which SKILL.md Is Good?
-
nanollama: Open-Source Framework for Training Llama 3 from Scratch with One-Command GGUF Export
-
CPU-Trained Language Model Outperforms GPU Baseline After 40 Hours
-
Why AI Models Fail at Iterative Reasoning and What Could Fix It
-
GLM-5 Technical Report: DSA Innovation Reduces Training and Inference Costs
-
Matmul-Free Language Model Trained on CPU in 1.2 Hours
-
Cohere Releases Tiny Aya: Efficient 3.3B Multilingual Model for 70+ Languages
-
GPT-OSS 120B Uncensored Model Released in Native MXFP4 Precision
-
Optimal llama.cpp Settings Found for Qwen3 Coder Next Loop Issues
-
Student Releases Dhi-5B: Multimodal Model Trained for Just $1,200