Yekun Chai

london.jpg

I study language models through the lens of compression. My work centers on large-scale pre-training, driven by one question: at what point does prediction become planning? I have contributed to ERNIE and StarCoder 2.

News

Aug 21, 2025 Four papers accepted to EMNLP 2025.
May 16, 2025 Curiosity-driven RLHF accepted to ACL 2025. code
Jan 23, 2025 MA-RLHF accepted to ICLR 2025. paper code

Latest Posts

Selected Publications

  1. MA-RLHF: Reinforcement Learning from Human Feedback with Macro Actions
    Yekun Chai*†Haoran Sun*^Huang FangShuohuan Wang, and 2 more authors
    In The Thirteenth International Conference on Learning Representations, 2025
  2. ICLR Spotlight
    Tool-Augmented Reward Modeling
    Lei Li*^Yekun Chai*†Shuohuan WangYu Sun, and 3 more authors
    In The Twelfth International Conference on Learning Representations(top 5%) , 2024
  3. ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for Programming Languages
    Yekun ChaiShuohuan Wang, Chao Pang, Yu Sun, and 2 more authors
    In Findings of the Association for Computational Linguistics: ACL 2023, Jul 2023