Mathy AI Substack
Subscribe
Sign in
Home
Notes
Archive
About
Latest
Top
Discussions
Finetuning is a Memory Wipe. This Is How You Stop It
Scaling Laws for Forgetting When Fine-Tuning Large Language Models, Mike’s Daily Paper: 19.08.25
15 hrs ago
•
Mike Erlihson, Mathy AI
1
Share this post
Mathy AI Substack
Finetuning is a Memory Wipe. This Is How You Stop It
Copy link
Facebook
Email
Notes
More
The End of Transformer Babysitting: Forging Stability Without the Hacks.
Mike’s Daily Paper: 19.08.25 - A New Foundation for Stable Transformers: Enforcing Lipschitz Bounds
Aug 20
•
Mike Erlihson, Mathy AI
Share this post
Mathy AI Substack
The End of Transformer Babysitting: Forging Stability Without the Hacks.
Copy link
Facebook
Email
Notes
More
When an LLM Stops Talking and Starts Deleting Files — Enter the Age of LAMs
Omri & Mike’s Daily Paper: 16.08.25, Large Action Models: From Inception to Implementation
Aug 17
•
Mike Erlihson, Mathy AI
and
Omri Sapir
3
Share this post
Mathy AI Substack
When an LLM Stops Talking and Starts Deleting Files — Enter the Age of LAMs
Copy link
Facebook
Email
Notes
More
The Rise of the AI-Augmented Generalist: Navigating the Deep Tech Landscape
The future won't be about who knows the most about one thing, but who can leverage AI to connect everything. But specialists will continue to be vital…
Aug 16
•
Mike Erlihson, Mathy AI
4
Share this post
Mathy AI Substack
The Rise of the AI-Augmented Generalist: Navigating the Deep Tech Landscape
Copy link
Facebook
Email
Notes
More
The Reasoning Illusion: FormulaOne Exposes the Algorithmic Blind Spot of LLMs
Mike's Daily Paper: 14.08.25, FormulaOne: Measuring the Depth of Algorithmic Reasoning Beyond Competitive Programming
Aug 14
•
Mike Erlihson, Mathy AI
2
Share this post
Mathy AI Substack
The Reasoning Illusion: FormulaOne Exposes the Algorithmic Blind Spot of LLMs
Copy link
Facebook
Email
Notes
More
We've Been Aligning LLMs All Wrong. The Solution is Deceptively Simple
Mike's Daily Paper: 13.08.25, Checklists Are Better Than Reward Models For Aligning Language Model
Aug 13
•
Mike Erlihson, Mathy AI
Share this post
Mathy AI Substack
We've Been Aligning LLMs All Wrong. The Solution is Deceptively Simple
Copy link
Facebook
Email
Notes
More
Beyond Quadratic: A Deep Dive into the Landscape of Efficient Attention
Mike’s Daily Paper: 08.08.25: Efficient Attention Mechanisms for Large Language Models: A Survey
Aug 9
•
Mike Erlihson, Mathy AI
3
Share this post
Mathy AI Substack
Beyond Quadratic: A Deep Dive into the Landscape of Efficient Attention
Copy link
Facebook
Email
Notes
More
The Glass Wall of Knowledge: Why Today's AI Can't Make True Discoveries
We haven't built (yet) an artificial scientist; we've just built a very expensive mirror that only reflects what we already know.
Aug 8
•
Mike Erlihson, Mathy AI
Share this post
Mathy AI Substack
The Glass Wall of Knowledge: Why Today's AI Can't Make True Discoveries
Copy link
Facebook
Email
Notes
More
It’s Not Just What You Prompt, It’s Where
Mike’s Daily Paper: 06.08.25 - Where to show Demos in Your Prompt: A Positional Bias of In-Context Learning
Aug 6
•
Mike Erlihson, Mathy AI
Share this post
Mathy AI Substack
It’s Not Just What You Prompt, It’s Where
Copy link
Facebook
Email
Notes
More
Physics Meets AI: How a New Model Learns Language Without Predicting a Single Token.
Mike’s Daily Paper: 04.08.25Rethinking Transformers Through the Lens of Physics: The Rise of Energy-Based Models
Aug 4
•
Mike Erlihson, Mathy AI
1
Share this post
Mathy AI Substack
Physics Meets AI: How a New Model Learns Language Without Predicting a Single Token.
Copy link
Facebook
Email
Notes
More
Do All Tokens Need the Same Amount of "Thinking"? Mixture-of-Recursions Says No.
Mike's Daily Paper: 02.08.25 - Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive Token-Level Computation
Aug 2
•
Mike Erlihson, Mathy AI
1
Share this post
Mathy AI Substack
Do All Tokens Need the Same Amount of "Thinking"? Mixture-of-Recursions Says No.
Copy link
Facebook
Email
Notes
More
Thinking Without Words: The Architectural Revolution(or not) AI Has Been Waiting For
Mike's Daily Paper: 01.08.25 - Hierarchical Reasoning Model
Aug 1
•
Mike Erlihson, Mathy AI
1
Share this post
Mathy AI Substack
Thinking Without Words: The Architectural Revolution(or not) AI Has Been Waiting For
Copy link
Facebook
Email
Notes
More
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts