Minsu Kim

profile_minsu.jpg

I’ll be joining Microsoft’s Copilot Tuning Research team as a Senior Researcher in 2026. There, I will work on advancing Copilot agents through research on LLM post-training, with a focus on reinforcement learning, agentic behavior, and reliable decision-making.

I am currently a postdoctoral fellow working with Prof. Yoshua Bengio at Mila, and Prof. Sungjin Ahn and Prof. Sungsoo Ahn at KAIST.

My research focuses on reinforcement learning, particularly on sample efficiency, credit assignment, and exploration, with applications to frontier LLMs, AI safety, and AI4Science.


Background

I received my Ph.D. from KAIST under the guidance of Prof. Jinkyoo Park, where I worked on bridging reinforcement learning and combinatorial optimization. My dissertation received the Presidential Best Ph.D. Thesis Award.

During my Ph.D., I collaborated with Prof. Sungsoo Ahn and conducted extended research visits at Mila, where I worked with Prof. Yoshua Bengio and his group on GFlowNet-related research.

I also received my master’s degree from KAIST under the supervision of Prof. Joungho Kim, where I worked on signal and power integrity in 2.5D/3D semiconductor systems, including HBM, using deep learning-based optimization methods.

Before that, I earned my B.S. in Mathematics and Computer Science from KAIST.

Academic Service

  • Area Chair: NeurIPS (Position Paper Track, 2026)
  • Reviewer (Conferences): NeurIPS (2022–2025), ICML (2023–2026), ICLR (2024–2026)
  • Reviewer (Journals): TNNLS (2025–2026), TPAMI (2025), TMLR (2025)

news

Feb 08, 2026 2 papers are accepted at ICLR 2026!
Sep 25, 2025 4 papers are accepted at NeurIPS 2025!
May 01, 2025 2 papers are accepted at ICML 2025!
Apr 01, 2025 I’ve selected Jang Yeong SIL Fellowship Award.
Feb 14, 2025 I got Ph.D degree with the KAIST presidential best Ph.D. thesis award.

latest posts

selected publications

  1. ICLR
    Latent Veracity Inference for Identifying Errors in Stepwise Reasoning
    Minsu Kim*, Jean-Pierre Falet*, Oliver E Richardson, Xiaoyin Chen, Moksh Jain, Sungjin Ahn, Sungsoo Ahn, and Yoshua Bengio
    International Conference on Learning Representations, 2026
  2. Thesis
    Off-policy Training Methods for Probablistic Agents in Combinatorial Space
    Minsu Kim
    Korea Advanced Institute of Science and Technology (KAIST), 2025
  3. AISTATS
    Ant Colony Sampling with GFlowNets for Combinatorial Optimization
    Minsu Kim*, Sanghyeok Choi*, Jiwoo Son, Hyeonah Kim, Jinkyoo Park, and Yoshua Bengio
    International Conference on Artificial Intelligence and Statistics, 2025
  4. ICLR
    Adaptive Teachers for Amortized Samplers
    Minsu Kim*, Sanghyeok Choi*, Taeyoung Yun, Emmanuel Bengio, Leo Feng, Jarrid Rector-Brooks, Sungsoo Ahn, Jinkyoo Park, Nikolay Malkin, and Yoshua Bengio
    International Conference on Learning Representations, 2025
  5. ICML
    Learning to Scale Logits for Temperature-Conditional GFlowNets
    Minsu Kim*, Joohwan Ko*, Taeyoung Yun*, Dinghuai Zhang, Ling Pan, Woochang Kim, Jinkyoo Park, Emmanuel Bengio, and Yoshua Bengio
    International Conference on Machine Learning, 2024
  6. ICLR
    Local Search GFlowNets
    Minsu Kim, Taeyoung Yun, Emmanuel Bengio, Dinghuai Zhang, Yoshua Bengio, Sungsoo Ahn, and Jinkyoo Park
    International Conference on Learning Representations, 2024
  7. NeurIPS
    Bootstrapped Training of Score-Conditioned Generator for Offline Design of Biological Sequences
    Minsu Kim, Federico Berto, Sungsoo Ahn, and Jinkyoo Park
    Advances in Neural Information Processing Systems, 2023
  8. NeurIPS
    Sym-NCO: Leveraging Symmetricity for Neural Combinatorial Optimization
    Minsu Kim, Junyoung Park, and Jinkyoo Park
    Advances in Neural Information Processing Systems, 2022
  9. NeurIPS
    Learning collaborative policies to solve NP-hard routing problems
    Minsu Kim, Jinkyoo Park, and Joungho Kim
    Advances in Neural Information Processing Systems, 2021