Investigation into Training Dynamics of Learned Optimizers (Student Abstract)

Abstract

Modern machine learning heavily relies on optimization, and as deep learning models grow more complex and data-hungry, the search for efficient learning becomes crucial. Learned optimizers disrupt traditional handcrafted methods such as SGD and Adam by learning the optimization strategy itself, potentially speeding up training. However, the learned optimizers’ dynamics are still not well understood. To remedy this, our work explores their optimization trajectories from the perspective of network architecture symmetries and proposed parameter update distributions.

Publication
The 38th Annual AAAI Conference on Artificial Intelligence (AAAI-24)

Published in the proceedings of the proceedings of the AAAI Conference on Artificial Intelligence, 38(21), 23657-23658. https://doi.org/10.1609/aaai.v38i21.30514.

Jan Sobotka
Jan Sobotka
CS Master’s Student & AI/ML Research Assistant

I am a master’s student in computer science at EPFL, and a research assistant at the MLBio Lab. I am interested in representation learning, (mechanistic) interpretability, meta-learning, reasoning, test-time training, and machine consciousness.