#modelcompression
FiD-GP reduces Bayesian training cost by orders of magnitude and halves parameter counts, shrinking model size by three-quarters, keeping state-of-the-art accuracy. Read more: https://getnews.me/flow-induced-diagonal-gaussian-processes-enhance-ai-model-compression/ #bayesiandeep #modelcompression
October 6, 2025 at 3:43 PM
In‑training compression trims SSM hidden dimensions during training, preserving performance while speeding up optimization; paper submitted Oct 2025. Read more: https://getnews.me/in-training-compression-improves-efficiency-of-state-space-models/ #statespacemodels #modelcompression
October 6, 2025 at 8:09 AM
Dynamic expert clustering cuts MoE model parameters by about 80% and boosts throughput 10‑20% while keeping quality on GLUE and WikiText‑103. https://getnews.me/dynamic-expert-clustering-boosts-efficiency-of-moe-large-language-models/ #moe #modelcompression #nlp
October 6, 2025 at 4:27 AM
BALF enables fine-tuning-free compression, cutting FLOPs of ResNeXt-101 by about 45% while incurring only a 1-point top-1 accuracy drop. The paper was submitted in September 2025. Read more: https://getnews.me/balf-enables-fine-tuning-free-neural-network-compression/ #balf #modelcompression
October 1, 2025 at 3:22 AM
RMT‑KD leverages random matrix theory for knowledge distillation, trimming up to 80% of model parameters with just ~2% accuracy loss and 2.8× faster inference. https://getnews.me/random-matrix-theory-powers-new-ai-model-compression-technique/ #randommatrixtheory #modelcompression
September 30, 2025 at 12:58 AM
COSPADI compresses large language models without additional training, using calibration‑guided sparse dictionary factorization to achieve 20‑50% reduction while preserving accuracy. https://getnews.me/cospadi-sparse-dictionary-learning-boosts-llm-compression/ #llm #modelcompression #sparselearning
September 29, 2025 at 11:36 AM
SlimDiff compresses diffusion models without training, achieving up to 35% faster inference and removing about 100 million parameters while maintaining quality. https://getnews.me/slimdiff-enables-training-free-compression-of-diffusion-models/ #slimdiff #diffusionmodels #modelcompression
September 29, 2025 at 5:50 AM
A unified framework merges tensor decomposition with automatic rank selection, cutting manual grid searches and using continuous optimization to compress models while keeping accuracy. https://getnews.me/unified-framework-for-neural-network-compression-with-rank-selection/ #modelcompression #nn
September 25, 2025 at 1:11 PM
Location‑aware discriminant analysis compresses detectors, cutting model size while preserving accuracy; on KITTI and COCO the pruned models matched or beat the originals. https://getnews.me/location-aware-discriminant-analysis-improves-visual-detector-compression/ #locationaware #modelcompression
September 25, 2025 at 5:08 AM
FiD‑GP halves neural network parameters and shrinks storage by about 75 %, while keeping state‑of‑the‑art accuracy and uncertainty estimation on benchmarks. Read more: https://getnews.me/flow-induced-diagonal-gaussian-processes-reduce-ai-model-size/ #fidgp #modelcompression
September 24, 2025 at 6:17 PM
RMT‑KD cuts model parameters by up to 80% with just 2% accuracy loss and runs up to 2.8× faster, tested on GLUE, AG News and CIFAR‑10, according to the study. Read more: https://getnews.me/random-matrix-theory-boosts-model-compression-with-rmt-kd/ #rmtkd #modelcompression #edgeai
September 22, 2025 at 11:19 AM
The new LWIQ method cuts tensor‑train rank‑search time by 63.2% and yields a model 3.2× smaller with only a 0.86% drop in top‑1 accuracy on CIFAR‑10 ResNet‑56. https://getnews.me/adaptive-tensor-train-decomposition-improves-network-compression/ #modelcompression #tensortrain #deeplearning
September 18, 2025 at 8:47 PM
ButterflyQuant slashes memory use in large language models without losing performance. Could this mean faster, cheaper AI on any device? What excites you about the future of model compression? 🤔 #AI #Innovation #ModelCompression LINK
September 14, 2025 at 2:42 AM
What is knowledge distillation and how does it work?

See here - techchilli.com/artificial-i...
June 24, 2025 at 6:24 AM
also provide a theoretical explanation for the observed gains. All codes of this paper are available at https://github.com/CSU-ModelCompression/BAQ. [7/7 of https://arxiv.org/abs/2506.05664v1]
June 9, 2025 at 6:05 AM
We propose Redundant Information Distillation which maximizes the task-relevant common information between teacher and student using a new alternating optimization: #explainability #informationtheory #distillation #modelcompression
May 2, 2025 at 10:45 PM
AI model compression isn't just a technical refinement but a strategic choice that aligns cost reduction, sustainability, and operational agility with the pressing demands of today's rapidly evolving digital landscape.

#AI #ModelCompression #Efficiency
March 22, 2025 at 2:30 PM
Today's task: model compression!!
🆕 New at IWSLT! But no less exciting 🔥

🎯 Goal: Compress a large, general-purpose multimodal model, making speech translation more efficient ⚡️, deployable 📲, and sustainable ♻️, while preserving translation quality ⭐️
#AI #SpeechTech #ModelCompression #LLMcompression
January 29, 2025 at 4:47 PM