BotBeat
...
← Back

> ▌

IntelIntel
RESEARCHIntel2026-04-14

MIT Researchers Develop CompreSSM: A Technique to Compress AI Models During Training Rather Than After

Key Takeaways

  • ▸CompreSSM compresses models during training rather than post-hoc, eliminating the traditional trade-off between model size and performance
  • ▸The technique uses control theory principles (Hankel singular values) to identify and rank component importance, with rankings stabilizing after just 10% of training
  • ▸Compressed models achieved up to 1.5x faster training on image classification and 4x speedups on Mamba architecture while maintaining competitive accuracy
Source:
Hacker Newshttps://news.mit.edu/2026/new-technique-makes-ai-models-leaner-faster-while-still-learning-0409↗

Summary

Researchers at MIT's Computer Science and Artificial Intelligence Laboratory, in collaboration with Max Planck Institute for Intelligent Systems, European Laboratory for Learning and Intelligent Systems, ETH, and Liquid AI, have developed CompreSSM, a novel technique that compresses AI models during the training process rather than after. The method targets state-space models used in language processing, audio generation, and robotics by employing mathematical tools from control theory to identify and remove unnecessary components early in training. Using Hankel singular values to measure the importance of internal states, the team demonstrated that component rankings stabilize after just 10 percent of training, allowing the remaining 90 percent to proceed at the speed of a much smaller model. The technique achieved striking results: compressed models maintained nearly identical accuracy to full-sized counterparts while training up to 1.5 times faster on image classification tasks, and achieved approximately 4x training speedups on Mamba architecture while reducing dimensionality from 128 to 12 dimensions.

  • This approach reduces computational resources, energy costs, and training time without requiring initial training of oversized models

Editorial Opinion

CompreSSM represents a paradigm shift in model optimization by integrating compression into the learning process itself rather than treating it as a post-hoc engineering problem. This work has significant implications for democratizing AI development, as it reduces the computational barriers to training performant models. The theoretical grounding using control theory provides a principled foundation that could inspire similar innovations across other model architectures beyond state-space models.

Machine LearningDeep LearningMLOps & InfrastructureScience & Research

More from Intel

IntelIntel
INDUSTRY REPORT

China Narrows AI Gap with US as Tech Talent Flow Slows, Stanford Report Finds

2026-04-17
IntelIntel
RESEARCH

Intel Arc Pro B70 Benchmarked Against AMD Radeon AI Pro and NVIDIA RTX on Linux

2026-04-16
IntelIntel
RESEARCH

Stanford Researchers Introduce TRACE: A System for Targeted Agent Self-Improvement Through Capability-Specific Training

2026-04-15

Comments

Suggested

OpenAIOpenAI
RESEARCH

OpenAI's GPT-5.4 Pro Solves Longstanding Erdős Math Problem, Reveals Novel Mathematical Connections

2026-04-17
CloudflareCloudflare
UPDATE

Cloudflare Enables AI-Generated Apps to Have Persistent Storage with Durable Objects in Dynamic Workers

2026-04-17
Institute for Basic Science (IBS)Institute for Basic Science (IBS)
RESEARCH

Scientists Develop SynTrogo Tool to Selectively Edit Brain Circuits and Enhance Memory

2026-04-17
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us