Yuekun Yao

About me

photo

Hi! I am Yuekun. I am currently a postdoctoral researcher at The Ohio State University, working with Prof. Huan Sun. I completed my Ph.D. in the Department of Language Science and Technology at Saarland University, where I worked with Prof. Alexander Koller. I was also part of the Computational Linguistics Group. In the past, I obtained my M.Sc. degree in Artificial Intelligence at the University of Edinburgh. Before that, I earned my B.S. in Computer Science at East China Normal University.

The main research question I am interested in is How does NLP models generalize to unfamiliar data and how can we improve it? I investigate out-of-distribution generalization with a focus on compositional generalization to bridge the gap between training and test distributions in realistic applications.

My work aims to both understand model behaviours and develop more effective and reliable NLP models through the following research questions.

  • Can NLP models or LLMs perform human-like generalization, and why? [1], [2] [3]
  • How to improve models' compositional generalization ability? [4]
  • How to build trustworthy models that generalize reliably? Can we train one model (discriminator) to judge the outputs of another model (parser)? [5]

Publications [Google Scholar][Semantic Scholar]

2025

Language models can learn implicit multi-hop reasoning, but only if they have lots of training data [paper]
Yuekun Yao, Yupei Du, Dawei Zhu, Michael Hahn, Alexander Koller
EMNLP 2025
Reason to rote: Rethinking memorization in reasoning [paper]
Yupei Du, Philipp Mondorf, Silvia Casola, Yuekun Yao, Robert Litschko, Barbara Plank
EMNLP 2025
Anything Goes? A Crosslinguistic Study of (Im)possible Language Learning in LMs [paper]
Xiulin Yang, Tatsuya Aoyama, Yuekun Yao, Ethan Wilcox
ACL 2025

2024

Predicting generalization performance with correctness discriminators [paper]
Yuekun Yao, Alexander Koller
EMNLP 2024 Findings
Simple and effective data augmentation for compositional generalization [paper] [code]
Yuekun Yao, Alexander Koller
NAACL 2024

2023

SLOG: A Structural Generalization Benchmark for Semantic Parsing [paper] [code]
Bingzhi Li, Lucia Donatelli, Alexander Koller, Tal Linzen, Yuekun Yao, Najoung Kim
EMNLP 2023

2022

Structural generalization is hard for sequence-to-sequence models [paper] [code] [data]
Yuekun Yao, Alexander Koller
EMNLP 2022

2020

Dynamic masking for improved stability in online spoken language translation [paper]
Yuekun Yao, Barry Haddow
The 14th Biennial Conference of the Association for Machine Translation in the Americas (AMTA 2020)
ELITR non-native speech translation at IWSLT 2020 [paper]
Dominik Macháček, Jonáš Kratochvíl, Sangeet Sagar, Matúš Žilinec, Ondřej Bojar, Thai-Son Nguyen, Felix Schneider, Philip Williams, Yuekun Yao
The 17th International Conference on Spoken Language Translation (IWSLT 2020)

Contact me

ykyao [dot] cs [at] gmail [dot] com