Discussion about this post

User's avatar
Neural Foundry's avatar

The idea that optimizaton algorithms like Adam are actually memory systms is such a fresh way to think about it. I've always thought of them as just tuning knobs but seeing them as information compression mechanisms realy changes the persective. If Google can make this work at scale without expensive retraining, it could finaly solve the continal learning problem that's been holding AI back?

Expand full comment

No posts

Ready for more?