Yafu Li (ζŽι›…ε€«)

I am a researcher at Shanghai AI Lab, working under the supervision of Prof. Yu Cheng. I earned my PhD through a joint program between Zhejiang University and Westlake University, advised by Prof. Yue Zhang.

I received my Bachelor's degree from Wuhan University, followed by a Master's degree from the University of Edinburgh, where I was supervised by Prof. Alex Lascarides. During my PhD, I interned at Tencent AI Lab and collaborated closely with Dr. Leyang Cui and Dr. Wei Bi. Prior to that, I worked as an NLP researcher at Huawei Noah's Ark Lab, under the mentorship of Dr. Liangyou Li and Prof. Qun Liu.

Email  /  Google Scholar  /  Twitter  /  Github

profile photo
Open Positions

We are looking for interns and joint PhD candidates (with THU, PKU, SJTU, FDU, etc.) to work on cutting-edge research in large language models. Our focus areas include zero reinforcement learning (e.g., R1-zero), test-time scaling, and trustworthy AI. If you are interested, please feel free to contact me at yafuly@gmail.com.

Research Areas

My research focuses on reasoning and inference scaling, trustworthy AI and machine translation. *: equal contributions. †: project lead or corresponding author.

Reasoning and Inference Scaling
PontTuset Scaling Reasoning, Losing Control: Evaluating Instruction Following in Large Reasoning Models
Tingchen Fu, Jiawei Gu, Yafu Li, Xiaoye Qu, Yu Cheng
preprint
Github / Paper

A benchmark to systematically evaluate and analyze the instruction-following ability of large reasoning models.

PontTuset Learning to Reason under Off-Policy Guidance
Jianhao Yan*, Yafu Li*, Zican Hu, Zhi Wang, Ganqu Cui, Xiaoye Qu, Yu Cheng, Yue Zhang
preprint
Github / Paper

A RL framework to boost reasoning performance using off-policy guidance.

PontTuset A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond
Xiaoye Qu*, Yafu Li*, Zhaochen Su, Weigao Sun, Jianhao Yan, Dongrui Liu, Ganqu Cui, Daizong Liu, Shuxian Liang, Junxian He, Peng Li, Wei Wei, Jing Shao, Chaochao Lu, Yue Zhang, Xian-Sheng Hua, Bowen Zhou, Yu Cheng
preprint
Github / Paper

We survey recent efforts to improve reasoning efficiency in Large Reasoning Models, highlighting inefficiency patterns in long reasoning traces and discussing solutions across the model lifecycle from pretraining to inference.

PontTuset Test-Time Preference Optimization: On-the-Fly Alignment via Iterative Textual Feedback
Yafu Li, Xuyang Hu, Xiaoye Qu, Linjie Li, Yu Cheng
ICML, 2025
Github / Paper

We present Test-Time Preference Optimization (TPO) that aligns LLMs during inference, surpassing strong baselines aligned during training.

PontTuset From Drafts to Answers: Unlocking LLM Potential via Aggregation Fine-Tuning
Yafu Li*, Zhilin Wang*, Tingchen Fu, Ganqu Cui, Sen Yang, Yu Cheng
preprint
Github / Paper

We present Aggregation Fine-Tuning (AFT) where the model learns to aggregate multiple drafts into a single answer.

PontTuset Multi-LLM Collaborative Search for Complex Problem Solving
Sen Yang, Yafu Li, Wai Lam, Yu Cheng
preprint
Paper

We propose Mixture-of-Search-Agents (MoSA), a test-time framework that coordinates multiple LLMs via Monte Carlo Tree Search to improve complex reasoning by aggregating diverse search trajectories.

Trustworthy AI
PontTuset Unveiling Attractor Cycles in Large Language Models: A Dynamical Systems View of Successive Paraphrasing
Zhilin Wang*, Yafu Li*, Jianhao Yan, Yu Cheng, Yue Zhang
ACL, 2025
Paper

LLMs exhibit stable paraphrasing cycles revealing limits in linguistic diversity through dynamical systems analysis.

PontTuset MAGE: Machine-generated Text Detection in the Wild
Yafu Li, Qintong Li, Leyang Cui, Wei Bi, Longyue Wang, Linyi Yang, Shuming Shi, Yue Zhang
ACL, 2024
Github / Paper

We present a comprehensive benchmark dataset designed to assess the proficiency of machine-generated text detectors amidst real-world scenarios.

PontTuset Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language Models
Yue Zhang, Yafu Li, Leyang Cui, Deng Cai, Lemao Liu, Tingchen Fu, Xinting Huang, Enbo Zhao, Yu Zhang, Yulong Chen, Longyue Wang, Anh Tuan Luu, Wei Bi, Freda Shi, Shuming Shi
Computational Linguistics
Github / Paper

A survey of hallucination in LLMs.

Machine Translation
PontTuset Lost in Literalism: How Supervised Training Shapes Translationese in LLMs
Yafu Li*, Ronghao Zhang*, Zhilin Wang, Huajiang Zhang, Leyang Cui, Yongjing Yin, Tong Xiao, Yue Zhang
ACL, 2025
Github / Paper

A systematic study of the origin of translationese in LLMs and mitigation methods.

PontTuset Explicit Syntactic Guidance for Neural Text Generation
Yafu Li, Leyang Cui, Jianhao Yan, Yongjing Yin, Wei Bi, Shuming Shi, Yue Zhang
ACL, 2023, Best Paper Nomination (1.6%)
Github / Paper

We propose a syntax-guided generation schema, which generates the sequence guided by a constituency parse tree in a top-down direction.

PontTuset Multi-Granularity Optimization for Non-Autoregressive Translation
Yafu Li, Leyang Cui, Yongjing Yin, Yue Zhang
EMNLP, 2022
Github / Paper

We propose multi-granularity optimization for non-autoregressive translation, which collects model behaviors on translation segments of various granularities and integrates feedback for backpropagation.

PontTuset On Compositional Generalization of Neural Machine Translation
Yafu Li, Yongjing Yin, Yulong Chen, Yue Zhang
ACL, 2021
Github / Paper

Neural machine translation suffers poor compositionality.

Education

PhD in Computer Science, Zhejiang University and Westlake University

Master of Science in Artificial Intelligence, University of Edinburgh

Bachelor of Engineering in Electronic Information Engineering, Wuhan University

Service

Area Chair: ACL 2025

Conference Reviewer: ACL, EMNLP, COLING, ACL ARR, IJCAI, NeurIPS.

Journal Reviewer: TMLR, JAIR, TACL, TASLP, TBD, TALLIP.

Honor

Outstanding Student Scholarship (Silver medal, Tencent Rhino-Bird Elite Program, 2024).

National Scholarship (Ministry of Education, 2023).

Dean's Medal (Westlake University, 2023).


Website's code is from Jon Barron.