Nov. 13, 2025

EP15: The Information Bottleneck and Scaling Laws with Alex Alemi

The player is loading ...
EP15: The Information Bottleneck and Scaling Laws with Alex Alemi

In this episode, we sit down with Alex Alemi, an AI researcher at Anthropic (previously at Google Brain and Disney), to explore the powerful framework of the information bottleneck and its profound implications for modern machine learning.

We break down what the information bottleneck really means, a principled approach to retaining only the most informative parts of data while compressing away the irrelevant. We discuss why compression is still important in our era of big data, how it prevents overfitting, and why it's essential for building models that generalize well.

We also dive into scaling laws: why they matter, what we can learn from them, and what they tell us about the future of AI research.

 

Papers and links:

 

Music:

“Kid Kodi” — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0.

“Palms Down” — Blue Dot Sessions — via Free Music Archive — CC BY-NC 4.0.

Changes: trimmed