Yekun Chai

Contact:
chaiyekun (at) gmail.com
I am a staff engineer at Baidu NLP, where I have contributed to Baidu’s LLM series such as ERNIE 5.0, 4.0, 3.5 and ERNIE-Code, and their industry-driven generative AI products such as ERNIE-Bot (文心一言) and Baidu Comate (文心快码). Before that, I honed my skills in RL and NLP at Institute of Automation, Chinese Academy of Sciences. I pursued my academic studies in NLP at University of Edinburgh, under the supervision of Adam Lopez and Naomi Saphra.
My research endeavors revolve around LLM scaling, with a particular emphasis on:
- Train-time scaling, including pre-training and post-training (esp. RL scaling), across languages, modalities, and tasks.
- Efficient alignment, reasoning, and inference at scale.
- Efficient multimodal tokenization.
news
Jan 23, 2025 |
One paper on MA-RLHF has been accepted to ICLR 2025![]() ![]() |
---|---|
Sep 21, 2024 |
Our papers on PixelGPT, GPTfluence, and TKEval have been accepted to EMNLP 2024 and Findings. ![]() |
May 02, 2024 |
One paper on GiLOT, an XAI approach for LLMs, has been accepted to ICML 2024. ![]() |
Feb 20, 2024 | One paper on HumanEval-XL, a multilingual code generation benchmark has been accepted to LREC-COLING 2024. We’ve released the code and data. |
Jan 16, 2024 |
One paper on reward models with tool-augmented feedback has been accepted to ICLR 2024 (spotlight)![]() ![]() |