Tagged "hugging-face"
- Hugging Face Releases One-Liner for Automatic Hardware Detection and Model Selection
- Mistral Small 4 119B Released with NVFP4 Quantisation Support
- OmniCoder-9B: Efficient Coding Model for 8GB GPUs
- Cicikus v3 Prometheus 4.4B – An Experimental Franken-Merge for Edge Reasoning
- Qwen 3.5 Derestricted Model Available for Local Deployment
- Sarvam AI Releases 30B and 105B Open-Source Models Trained from Scratch
- Local LLM Performance Improvements: A Year of Progress Since DeepSeek R1 Moment
- Open-Source llama.cpp Finds Long-Term Home at Hugging Face
- O-TITANS: Orthogonal LoRA Framework for Gemma 3 with Google TITANS Memory Architecture
- GGML Joins Hugging Face: What This Means for Local Model Optimization
- Open-Source + AI: ggml Joins Hugging Face, llama.cpp Stays Open—Local AI's Long-Term Home
- GGML.AI Acquired by Hugging Face
- Matmul-Free Language Model Trained on CPU in 1.2 Hours
- Qwen 3.5-397B-A17B Now Available for Local Inference with Aggressive Quantisation
- GPT-OSS 20B Now Runs 100% Locally in Browser via WebGPU
- MiniMax M2.5: 230B Parameter MoE Model Coming to HuggingFace