Tagged "model-compression"
- Mirai Announces $10M to Advance On-Device AI Performance for Consumer Devices
- Kioxia Sampling UFS 5.0 Embedded Flash Memory for Next-Generation Mobile Applications
- Enhanced Interface Speed Enables High-Performance On-Device AI Features in Smartphones
- At India AI Impact Summit, Intel Showcases AI PCs and Cost-Efficient Frugal AI
- Taalas Etches AI Models onto Transistors to Rocket Boost Inference
- Qwen3 Coder Next Remains Effective at Aggressive Quantization Levels
- Sarvam Brings AI to Feature Phones, Cars, and Smart Glasses
- Enhanced Quantization Visualization Methods for Understanding LLM Compression Trade-offs
- Kitten TTS V0.8 Released: State-of-the-Art Super-Tiny Text-to-Speech Model Under 25MB
- Samsung's REAM: Alternative Model Compression Technique