Noah Lee

Noah.jpg

I am an LLM Researcher of the Kanana Team of Kakao, where I work on post-training LLMs. Previously, I graduated my Master’s degree at the Kim Jaechul Graduate School of AI of KAIST, jointly advised by James Thorne and Jinwoo Shin.

My current main research interest lies in (but not confined to):

  • Bettering human representation of LLMs
  • Incuding test-time alignemnt and/or controllability of LLMs
  • Personalization & customization of LLMs

Feel free to get in touch!


News

Sep 15, 2025 I joined the Kanana Team of Kakao Corp. to work on LLM post-training.
Jul 1, 2025 Two papers (Robust RM & AlphaPO) have been accepted to ICML 2025!
Jan 24, 2025 Our RM cross-lingual paper has been accepted to NAACL 2025!
Dec 2, 2024 A paper has been accepted to COLING 2025
Oct 30, 2024 Check out our new preprint on cross-lingual transfer of reward models!

Publications

  1. alphapo.png
    AlphaPO - Reward Shape Matters for LLM Alignment
    Aman Gupta, Shao Tang, Qingquan Song, and 8 more authors
    ICML, 2025
  2. robustrm.png
    On the Robustness of Reward Models for Language Model Alignment
    Jiwoo Hong, Noah Lee, Eunki Kim, and 5 more authors
    ICML, 2025
  3. cross.png
    Cross-lingual Transfer of Reward Models in Multilingual Alignment
    Jiwoo Hong*, Noah Lee*, Rodrigo Martínez-Castaño, and 2 more authors
    NAACL, 2025
  4. biggen.png
    The BiGGen Bench: A Principled Benchmark for Fine-grained Evaluation of Language Models with Language Models
    Seungone Kim, Juyoung Suk, Ji Yong Cho, and 29 more authors
    NAACL (Best Paper), 2025
  5. conseval.png
    Evaluating the Consistency of LLM Evaluators
    Noah Lee*, Jiwoo Hong*, and James Thorne
    COLING, 2025
  6. orpo.png
    ORPO: Monolithic Preference Optimization without Reference Model
    Jiwoo Hong, Noah Lee, and James Thorne
    EMNLP, 2024
  7. mapo.png
    Margin-aware Preference Optimization for Aligning Diffusion Models without Reference
    Jiwoo Hong*, Sayak Paul*, Noah Lee, and 3 more authors
    Preprint, 2024
  8. humllm.png
    Can Large Language Models Capture Dissenting Human Voices?
    Noah Lee*, Na Min An*, and James Thorne
    EMNLP, 2023