Abstract: The talk introduces Any Compression via Iterative Pruning (ACIP), a novel approach designed to give users intuitive control over the compression-performance trade-off. ACIP uses a single gradient descent run of iterative pruning to establish a global parameter ranking, enabling immediate materialization of models of any target size. It demonstrates strong predictive performance on downstream tasks without costly fine-tuning and achieves state-of-the-art compression for open-weight LLMs, often complementing common quantization techniques.
talk-data.com
Topic
foundation models
1
tagged
Activity Trend
1
peak/qtr
2020-Q1
2026-Q1
Top Events
Virtual event: "Time Series Mastery"
1
@Snapchat: WWDC 2025 Lightning Talks ⚡️
1
AI Meetup (September): AI, GenAI and LLMs
1
#16: Compressing Foundation Models as Easy as Image Compression? by M. Genzel
1
Mastering Prompt Engineering: Unlock the Power of AI
1
AI-Powered Data & Search: Unlocking Intelligence Across Systems
1
August Members Talk Evening
1
Filtering by:
#16: Compressing Foundation Models as Easy as Image Compression? by M. Genzel
×