About Me
Hi! My name is Minhak Song, and I am an undergraduate student at KAIST, majoring in Mathematical Sciences (minor in Industrial and Systems Engineering). My research interests center on the foundations of modern machine learning, spanning the theory of deep learning, language models, generative models, and interactive decision making, with the goal of bridging theory and practice.
I have recently been studying the training dynamics of optimization algorithms in deep learning, advised by Prof. Chulhee Yun at KAIST AI and closely collaborating with Kwangjun Ahn at Microsoft Research. I was also a visiting student researcher at the University of Washington, hosted by Prof. Simon Shaolei Du, where I worked on reinforcement learning from human feedback (RLHF) from an optimization perspective. Starting this summer, I am working with Prof. Sewoong Oh at the University of Washington on zeroth-order optimization.
I’m always happy to discuss research and potential collaborations. Feel free to reach out!
Research Interests
- DL/RL/LLM Theory
- Optimization
News
- [Jan. 2026] Our paper on the implicit bias of per-sample Adam on separable data is accepted to ICLR 2026.
- [Dec. 2025] Our paper on the theory of the spurious alignment of SGD in ill-conditioned high-dimensional quadratics is accepted to ALT 2026.
- [Oct. 2025] I was selected as a Top Reviewer (top 8% of reviewers) at NeurIPS 2025.
- [Sep. 2025] Our paper on understanding the benefit of Schedule-Free Optimizer through the river-valley loss landscape is accepted to NeurIPS 2025.
- [Jun. 2025] I joined Prof. Sewoong Oh’s group as a visiting student researcher at the University of Washington.
- [May. 2025] Our paper on how the dataset, network architecture, and optimizer influence progressive sharpening is accepted to ICML 2025.
- [Jan. 2025] Our paper on SGD dynamics in an ill-conditioned valley (a.k.a. river-valley) loss landscape is accepted to ICLR 2025.
- [Jan. 2025] I joined Prof. Simon Shaolei Du’s group as a visiting student researcher at the University of Washington.
- [Jan. 2024] Our paper on the optimization characteristics of linear Transformers is accepted to ICLR 2024.
- [Sep. 2023] Our paper on understanding the Edge of Stability phenomenon in deep learning is accepted to NeurIPS 2023.
Publications
(* denotes equal contribution)
-
Ruizhe Shi*, Minhak Song*, Runlong Zhou, Zihan Zhang, Maryam Fazel, Simon S. Du
Manuscript
-
Beomhan Baek*, Minhak Song*, Chulhee Yun
International Conference on Learning Representations (ICLR) 2026
NeurIPS 2025 Workshop on Optimization for Machine Learning
-
Shenyang Deng, Boyao Liao, Zhuoli Ouyang, Tianyu Pang, Minhak Song, Yaoqing Yang
International Conference on Algorithmic Learning Theory (ALT) 2026
-
Minhak Song*, Beomhan Baek*, Kwangjun Ahn, Chulhee Yun
Neural Information Processing Systems (NeurIPS) 2025
ICML 2025 Workshop on High-dimensional Learning Dynamics
-
Geonhui Yoo, Minhak Song, Chulhee Yun
International Conference on Machine Learning (ICML) 2025
-
Minhak Song, Kwangjun Ahn, Chulhee Yun
International Conference on Learning Representations (ICLR) 2025
ICML 2024 Workshop on High-dimensional Learning Dynamics
-
Kwangjun Ahn*, Xiang Cheng*, Minhak Song*, Chulhee Yun, Ali Jadbabaie, Suvrit Sra
International Conference on Learning Representations (ICLR) 2024
NeurIPS 2023 Workshop on Mathematics of Modern Machine Learning (Oral)
-
Minhak Song, Chulhee Yun
Neural Information Processing Systems (NeurIPS) 2023
Services
Conference/Workshop Reviewer