About Me
Hi! I am an undergraduate student majoring in Mathematical Sciences at KAIST, where I am fortunate to be advised by Chulhee Yun. I was also a visiting student at the University of Washington, working with Simon Shaolei Du and Sewoong Oh.
My research interests broadly span the foundations of deep learning, with the goal of bridging theory and practice. Recently, I have been focusing on understanding the optimization dynamics in deep learning, particularly in the pre-training and post-training of language models, and leveraging these insights to design principled and efficient optimization algorithms.
Update: Starting in Fall 2026, I will join Stanford University as a PhD student in Computer Science, supported by the Stanford School of Engineering Fellowship.
Research Interests
- DL/RL/LLM Theory
- Optimization
News
- [Apr. 2026] Two papers (Zeroth-Order Edge of Stability, Dichotomy of RLHF and DPO) are accepted to ICML 2026.
- [Apr. 2026] I am giving a contributed talk on Zeroth-Order Edge of Stability at the ICLR 2026 Workshop on Scientific Methods for Understanding Deep Learning in Rio, Brazil.
- [Feb. 2026] Our paper won the Best Student Paper Award at ALT 2026.
- [Jan. 2026] Our paper on the implicit bias of per-sample Adam on separable data is accepted to ICLR 2026.
- [Dec. 2025] Our paper on the theory of the spurious alignment of SGD in ill-conditioned high-dimensional quadratics is accepted to ALT 2026.
- [Oct. 2025] I was selected as a Top Reviewer (top 8% of reviewers) at NeurIPS 2025.
- [Sep. 2025] Our paper on understanding the benefit of Schedule-Free Optimizer through the river-valley loss landscape is accepted to NeurIPS 2025.
- [Jun. 2025] I joined Sewoong Oh’s group as a visiting student researcher at the University of Washington.
- [May. 2025] Our paper on how the datasets, network architectures, and optimizers influence progressive sharpening is accepted to ICML 2025.
- [Jan. 2025] Our paper on identifying the spurious alignment of SGD in an ill-conditioned valley (a.k.a. river-valley) loss landscape is accepted to ICLR 2025.
- [Jan. 2025] I joined Simon Shaolei Du’s group as a visiting student researcher at the University of Washington.
- [Jan. 2024] Our paper on the optimization characteristics of linear Transformers is accepted to ICLR 2024.
- [Sep. 2023] Our paper on understanding the Edge of Stability in deep learning is accepted to NeurIPS 2023.
Publications
(* denotes equal contribution)
-
Minhak Song, Liang Zhang, Bingcong Li, Niao He, Michael Muehlebach, Sewoong Oh
International Conference on Machine Learning (ICML) 2026
ICLR 2026 Workshop on Scientific Methods for Understanding Deep Learning (Oral)
-
Ruizhe Shi*, Minhak Song*, Runlong Zhou, Zihan Zhang, Maryam Fazel, Simon S. Du
International Conference on Machine Learning (ICML) 2026
-
Beomhan Baek*, Minhak Song*, Chulhee Yun
International Conference on Learning Representations (ICLR) 2026
NeurIPS 2025 Workshop on Optimization for Machine Learning
-
Shenyang Deng, Boyao Liao, Zhuoli Ouyang, Tianyu Pang, Minhak Song, Yaoqing Yang
International Conference on Algorithmic Learning Theory (ALT) 2026 (Best Student Paper)
-
Minhak Song*, Beomhan Baek*, Kwangjun Ahn, Chulhee Yun
Neural Information Processing Systems (NeurIPS) 2025
ICML 2025 Workshop on High-dimensional Learning Dynamics
-
Geonhui Yoo, Minhak Song, Chulhee Yun
International Conference on Machine Learning (ICML) 2025
-
Minhak Song, Kwangjun Ahn, Chulhee Yun
International Conference on Learning Representations (ICLR) 2025
ICML 2024 Workshop on High-dimensional Learning Dynamics
-
Kwangjun Ahn*, Xiang Cheng*, Minhak Song*, Chulhee Yun, Ali Jadbabaie, Suvrit Sra
International Conference on Learning Representations (ICLR) 2024
NeurIPS 2023 Workshop on Mathematics of Modern Machine Learning (Oral)
-
Minhak Song, Chulhee Yun
Neural Information Processing Systems (NeurIPS) 2023
Services
Conference/Workshop Reviewer