Filter

Identifiable Token Correspondence for World Models

Identifiable Token Correspondence for World Models

Youngin Kim, Ray Sun, Inho Kim, Bumsoo Park, Hyun Oh Song

Language Model ICML 2026
How to Correctly Report LLM-as-a-Judge Evaluations

How to Correctly Report LLM-as-a-Judge Evaluations

Chungpa Lee, Thomas Zeng, Jongwon Jeong, Jy-yong Sohn, Kangwook Lee

Language Model ICML 2026
ReJump: A Tree-Jump Representation for Analyzing and Improving LLM Reasoning

ReJump: A Tree-Jump Representation for Analyzing and Improving LLM Reasoning

Yuchen Zeng, Shuibai Zhang, Wonjun Kang, Shutong Wu, Lynnix Zou, Ying Fan, Heeju Kim, Ziqian Lin, Jungtaek Kim, Hyung Il Koo, Dimitris Papailiopoulos, Kangwook Lee

Language Model ICML 2026
Fine-Tuning Without Forgetting In-Context Learning: A Theoretical Analysis of Linear Attention Models

Fine-Tuning Without Forgetting In-Context Learning: A Theoretical Analysis of Linear Attention Models

Chungpa Lee, Jy-yong Sohn, Kangwook Lee

Theoretical ICML 2026
Mitigating Perceptual Judgment Bias in Multimodal LLM-as-a-Judge via Perceptual Perturbation and Reward Modeling

Mitigating Perceptual Judgment Bias in Multimodal LLM-as-a-Judge via Perceptual Perturbation and Reward Modeling

Seojeong Park*, Jiho Choi*, Junyong Kang, Seonho Lee, Jaeyo Shin, Hyunjung Shim

Language Model ICML 2026
Understanding the Performance Gap in Preference Learning: A Dichotomy of RLHF and DPO

Understanding the Performance Gap in Preference Learning: A Dichotomy of RLHF and DPO

Ruizhe Shi*, Minhak Song*, Runlong Zhou, Zihan Zhang, Maryam Fazel, Simon S. Du

Language Model ICML 2026
Coverage Improvement and Fast Convergence of On-policy Preference Learning

Coverage Improvement and Fast Convergence of On-policy Preference Learning

Juno Kim, Jihun Yun, Jason D. Lee, Kwang-Sung Jun

Language Model ICML 2026
Zeroth-Order Optimization at the Edge of Stability

Zeroth-Order Optimization at the Edge of Stability

Minhak Song, Liang Zhang, Bingcong Li, Niao He, Michael Muehlebach, Sewoong Oh

Language Model ICML 2026
Lookahead Unmasking Elicits Accurate Decoding in Diffusion Language Models

Lookahead Unmasking Elicits Accurate Decoding in Diffusion Language Models

Sanghyun Lee, Seungryong Kim, Jongho Park, Dongmin Park

Language Model ICML 2026
Can Large Language Models Keep Up? Benchmarking Online Adaptation to Continual Knowledge Streams

Can Large Language Models Keep Up? Benchmarking Online Adaptation to Continual Knowledge Streams

Jiyeon Kim, Hyunji Lee, Dylan Zhou, Sue Hyun Park, Seunghyun Yoon, Trung Bui, Franck Dernoncourt, Sungmin Cha, Minjoon Seo

Language Model ACL 2026