publications

2025

  1. alphapo.png
    AlphaPO - Reward Shape Matters for LLM Alignment
    Aman Gupta, Shao Tang, Qingquan Song, and 8 more authors
    ICML, 2025
  2. robustrm.png
    On the Robustness of Reward Models for Language Model Alignment
    Jiwoo Hong, Noah Lee, Eunki Kim, and 5 more authors
    ICML, 2025
  3. cross.png
    Cross-lingual Transfer of Reward Models in Multilingual Alignment
    Jiwoo Hong*, Noah Lee*, Rodrigo Martínez-Castaño, and 2 more authors
    NAACL, 2025
  4. biggen.png
    The BiGGen Bench: A Principled Benchmark for Fine-grained Evaluation of Language Models with Language Models
    Seungone Kim, Juyoung Suk, Ji Yong Cho, and 29 more authors
    NAACL (Best Paper), 2025
  5. conseval.png
    Evaluating the Consistency of LLM Evaluators
    Noah Lee*, Jiwoo Hong*, and James Thorne
    COLING, 2025

2024

  1. orpo.png
    ORPO: Monolithic Preference Optimization without Reference Model
    Jiwoo Hong, Noah Lee, and James Thorne
    EMNLP, 2024
  2. mapo.png
    Margin-aware Preference Optimization for Aligning Diffusion Models without Reference
    Jiwoo Hong*, Sayak Paul*, Noah Lee, and 3 more authors
    Preprint, 2024

2023

  1. humllm.png
    Can Large Language Models Capture Dissenting Human Voices?
    Noah Lee*, Na Min An*, and James Thorne
    EMNLP, 2023