About
Hi! I'm Hyeonbin.
I’m a Master’s student at KAIST AI advised by Minjoon Seo.
My graduate research began with a simple question: How do language models reason, and how can we make them reason better? Along the way, I quickly ran into what Sutton calls the "bitter lesson": models often appear capable by memorizing patterns in the training data, only to plateau once they encounter situations outside that distribution.
Nevertheless, I study how to design learning signals that encourage models to move past surface-level pattern matching, and I aim to explore these capabilities through diverse domains. Currently, I am an intern at ALIN Lab, where I work on robotics + reinforcement learning.
I’m always open to collaborations—feel free to reach out!
Experience
-
ALIN LabKAISTResearch Intern (Current) 2025-Advisor: Jinwoo Shin
Robotics + RL -
KRAFTONAI Companion TeamInternship 2025
-
KAIST AIM.S. 2024 ~Advisor: Minjoon Seo
Language Model Reasoning -
NAVERHealthcare AIInternship 2022EHR summarization
-
KAIST SoCB.S. 2019 — 2024GPA 3.96 / 4.3
Publications
Selected Publications
-
Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks
H Hwang, Y Park Preprint (under review)
-
Let's Predict Sentence by Sentence
H Hwang, B Jeon, S Kim, J Kim, H Chang, S Yang, S Won, D Lee, Y Ahn, ... @ COLM 2025 RAM 2 WS (Oral)
-
Self-Explore to Avoid the Pit: Improving the Reasoning Capabilities of Language Models
H Hwang, D Kim, S Kim, S Ye, M Seo @ EMNLP 2024 (Findings)
@ ACL 2024 NLRSE WS (Oral)
Others
-
The Coverage Principle: A Framework for Understanding Compositional Generalization
H Chang, J Park, H Cho, S Yang, M Ko, H Hwang, S Won, D Lee, Y Ahn, ... @ ICLR 2026
-
The CoT Encyclopedia: Analyzing, Predicting, and Controlling how a Reasoning Model will Think
S Lee, S Kim, M Seo, Y Jo, D Go, H Hwang, J Park, X Yue, S Welleck, G Neubig, M Lee, M Seo @ ICLR 2026
-
Differential Information Distribution: A Bayesian Perspective on Direct Preference Optimization
Y Won, H Lee, H Hwang, M Seo Preprint
-
BiGGen Bench: A Comprehensive Benchmark for Generative Language Models
S Kim, J Suk, ... (others not shown), H Hwang, ... M Seo @ NAACL 2025 (Best Paper)
-
Knowledge Entropy Decay during Language Model Pretraining Hinders New Knowledge Acquisition
J Kim, H Lee, H Cho, J Jang, H Hwang, S Won, Y Ahn, D Lee, M Seo @ ICLR 2025 (Oral)
@ AAAI 2025 KnowFM WS (Best Paper)
-
FLASK: Fine-grained Language Model Evaluation Based on Alignment Skill Sets
S Ye, D Kim, S Kim, H Hwang, S Kim, Y Jo, J Thorne, J Kim, M Seo @ ICLR 2024 (Spotlight)