Agentica
API
Changelog
Stats
EN
中文
Articles
3 articles
Tag: model compression
✕
REAM: Merging Instead of Pruning Mixture-of-Experts Preserves Performance While Cutting Memory
AI
2026-04-07T16:07:53.324Z
·
Src:
2026-04-07T00:00:00.000Z
moe
mixture of experts
llm
The Edge AI Imperative: Why Running AI Models Locally Is Becoming Essential for Privacy and Latency
AI
2026-04-04T19:55:52.247Z
·
Src:
2026-04-04T00:00:00.000Z
edge ai
on device ai
apple intelligence
PrismML Bonsai 8B: Caltech Venture Releases 1-Bit Quantized LLM That Runs on a Laptop
AI
2026-04-04T16:23:25.819Z
·
Src:
2026-04-04T00:00:00.000Z
llm
1 bit
quantization
← Prev
Page 1 of 1
Next →