Tagged "model-compression"
- LLM Neuroanatomy II: Modern LLM Hacking and Hints of a Universal Language
- Running an AI Agent on a 448KB RAM Microcontroller
- Multiverse Computing Targets On-Device AI With Compressed Models and New API Portal
- Researcher Discovers Universal "Danger Zone" in Transformer Model Architecture at 50% Depth
- Nota Added to Three Technology and Growth ETFs in a Row – Market Recognition for AI Efficiency
- Nota AI to Showcase End-to-End On-Device AI Optimization at Embedded World 2026
- Student Researcher Achieves 42x Model Compression Through Novel Architecture
- ETH Zurich Research Challenges Context-Length Assumptions in LLM Agents
- OPPO and MediaTek Highlight On-Device AI Innovations at MWC 2026
- Qualcomm Snapdragon Wear Elite Brings On-Device AI to Smartwatches
- On-Device AI Laptop Lineups Become Standard Across Major Manufacturers
- Meta Reveals AI-Packed Smartwatch In 2026 – Why Wearables Shift Now
- Arduino and Qualcomm Bring On-Device AI Learning to Indian Schools
- Mirai Announces $10M to Advance On-Device AI Performance for Consumer Devices
- Kioxia Sampling UFS 5.0 Embedded Flash Memory for Next-Generation Mobile Applications
- Enhanced Interface Speed Enables High-Performance On-Device AI Features in Smartphones
- At India AI Impact Summit, Intel Showcases AI PCs and Cost-Efficient Frugal AI
- Sarvam Brings AI to Feature Phones, Cars, and Smart Glasses
- NVIDIA's Dynamic Memory Sparsification Cuts LLM Inference Costs by 8x
- Samsung's REAM: Alternative Model Compression Technique