avatar

Tianyu Liu

Researcher
Moonshot AI (Kimi)



Biography

Tianyu Liu is a researcher at Kimi working on coding and agents, and leads an effort toward developing foundational models for coding agents and broader agentic experiences. Before joining Kimi, he was a staff researcher in the early Qwen team at Alibaba, focusing on reasoning and coding, and prior to that a senior researcher and founding member of Tencent Hunyuan. He received his PhD from Peking University in 2021, advised by Zhifang Sui and Baobao Chang. During his PhD, he also had the opportunity to intern at or visit Microsoft Research (Beijing and Redmond) and Toyota Technological Institute at Chicago (TTIC).

Before the rise of LLMs, Tianyu's research centered on natural language generation, information extraction, and robustness in NLP. The arrival of large language models reshaped his trajectory: starting in mid 2022 at Tencent, he co-led the development of coding-oriented models and internal Copilot-style systems, and later became a founding member of Hunyuan. He then joined the early Qwen team at Alibaba, where he contributed across pretraining, mid-training, and post-training for reasoning and coding — serving as a core contributor to Qwen-Math and Qwen-Coder. He also led a team that won a gold medal at AIMO-2.

Now at Kimi, his work covers the full stack of coding and agent capability — from pretraining data and long-CoT reasoning to RL-based training, coding environments, and the supporting infrastructure. These efforts have contributed to a series of Kimi models such as K2.5. More broadly, he is passionate about building foundation models that can see, reason, code, and act — combining multimodal perception with long-horizon planning to deliver truly reliable and practical agent experiences.

Selected Publications [Google Scholar][DBLP]

* indicates equal contribution, indicates project lead or corresponding author

Kimi-Dev: Agentless Training as Skill Prior for SWE-Agents
Zonghan Yang, Shengjie Wang, Kelin Fu, Wenyang He, Weimin Xiong, Yibo Liu, Yibo Miao, Bofei Gao, Yejie Wang, Yingwei Ma, Yanhao Li, Yue Liu, Zhenxing Hu, Kaitai Zhang, Shuyi Wang, Huarong Chen, Flood Sung, Yang Liu, Yang Gao, Zhilin Yang, Tianyu Liu
ICLR 2026
[PDF] [bib]
Multi-Docker-Eval: A 'Shovel of the Gold Rush' Benchmark on Automatic Environment Building for Software Engineering
Kelin Fu, Tianyu Liu, Zeyu Shang, Yingwei Ma, Jian Yang, Jiaheng Liu, Kaigui Bian
[PDF] [bib]
OJBench: A Competition Level Code Benchmark for Large Language Models
Zhexu Wang, Yiping Liu, Yejie Wang, Wenyang He, Bofei Gao, Muxi Diao, Yanxu Chen, Kelin Fu, Flood Sung, Zhilin Yang, Tianyu Liu, Weiran Xu
[PDF] [bib]
LLM Critics Help Catch Bugs in Mathematics: Towards a Better Mathematical Verifier with Natural Language Feedback
Bofei Gao, Zefan Cai, Runxin Xu, Peiyi Wang, Ce Zheng, Runji Lin, Keming Lu, Dayiheng Liu, Chang Zhou, Wen Xiao, Junjie Hu, Tianyu Liu, Baobao Chang
ACL 2025 Findings
[PDF] [bib]
Omni-MATH: A Universal Olympiad Level Mathematic Benchmark for Large Language Models
Bofei Gao, Feifan Song, Zhe Yang, Zefan Cai, Yibo Miao, Qingxiu Dong, Lei Li, Chenghao Ma, Liang Chen, Runxin Xu, Zhengyang Tang, Benyou Wang, Daoguang Zan, Shanghaoran Quan, Ge Zhang, Lei Sha, Yichang Zhang, Xuancheng Ren, Tianyu Liu, Baobao Chang
[PDF] [bib]
An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models
Liang Chen, Haozhe Zhao, Tianyu Liu, Shuai Bai, Junyang Lin, Chang Zhou, Baobao Chang
ECCV 2024
[PDF] [bib] [Code]
Making Large Language Models Better Reasoners with Alignment
Peiyi Wang, Lei Li, Liang Chen, Feifan Song, Binghuai Lin, Yunbo Cao, Tianyu Liu, Zhifang Sui
EMNLP 2024 Findings
[PDF] [bib]
Large Language Models are not Fair Evaluators
Peiyi Wang, Lei Li, Liang Chen, Zefan Cai, Dawei Zhu, Binghuai Lin, Yunbo Cao, Qi Liu, Tianyu Liu, Zhifang Sui
ACL 2024
[PDF] [bib] [FairEval]
Unlocking Efficiency in Large Language Model Inference: A Comprehensive Survey of Speculative Decoding
Heming Xia, Zhe Yang, Qingxiu Dong, Peiyi Wang, Yongqi Li, Tao Ge, Tianyu Liu, Wenjie Li, Zhifang Sui
ACL 2024 Findings
[PDF] [bib] [Spec-Bench] [Paper List]
Towards End-to-End Embodied Decision Making via Multi-modal Large Language Model: Explorations with GPT4-Vision and Beyond
Liang Chen, Yichi Zhang, Shuhuai Ren, Haozhe Zhao, Zefan Cai, Yuchi Wang, Peiyi Wang, Tianyu Liu, Baobao Chang
NeurIPS 2023 FMDM Workshop (Oral)
[PDF] [bib] [PCA-Eval]
DialogVCS: Robust Natural Language Understanding in Dialogue System Upgrade
Zefan Cai*, Xin Zheng*, Tianyu Liu*, Xu Wang, Haoran Meng, Jiaqi Han, Gang Yuan, Binghuai Lin, Baobao Chang, Yunbo Cao
[PDF] [bib]
A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation
Tianyu Liu, Yizhe Zhang, Chris Brockett, Yi Mao, Zhifang Sui, Weizhu Chen, Bill Dolan
Association for Computational Linguistics (ACL), 2022
[PDF] [bib] [code]
DialogUSR: Complex Dialogue Utterance Splitting and Reformulation for Multiple Intent Detection
Haoran Meng*, Zheng Xin*, Tianyu Liu*, Zizhen Wang, He Feng, Binghuai Lin, Xuemin Zhao, Yunbo Cao, Zhifang Sui
Findings of Empirical Methods in Natural Language Processing (Findings of EMNLP), 2022
[PDF] [bib] [code]
An Enhanced Span-based Decomposition Method for Few-Shot Sequence Labeling
Peiyi Wang*, Runxin Xu*, Tianyu Liu, Qingyu Zhou, Yunbo Cao, Baobao Chang, Zhifang Sui
North American Association for Computational Linguistics (NAACL), 2022
[PDF] [bib] [code]
Towards Faithfulness in Open Domain Table-to-text Generation from an Entity-centric View
Tianyu Liu*, Xin Zheng*, Baobao Chang, Zhifang Sui
AAAI Conference on Artificial Intelligence (AAAI), 2021
[PDF] [bib]
An Anchor-Based Automatic Evaluation Metric for Document Summarization
Kexiang Wang*, Tianyu Liu*, Baobao Chang, Zhifang Sui
International Conference on Computational Linguistics (COLING), 2020
[PDF] [bib]
An Empirical Study on Model-agnostic Debiasing Strategies for Robust Natural Language Inference
Tianyu Liu*, Xin Zheng*, Xiaoan Ding, Baobao Chang, Zhifang Sui
Conference on Computational Natural Language Learning (CoNLL), 2020
[PDF] [bib] [Adversarial Datasets]
Discrimatively-Tuned Generative Classifier for Robust Natural Language Inference
Xiaoan Ding*, Tianyu Liu*, Baobao Chang, Zhifang Sui, Kevin Gimpel
Empirical Methods in Natural Language Processing (EMNLP), 2020
[PDF] [bib]
An Exploration of Arbitrary-Order Sequence Labeling via Energy-Based Inference Networks
Lifu Tu*, Tianyu Liu*, Kevin Gimpel
Empirical Methods in Natural Language Processing (EMNLP), 2020
[PDF] [bib] [code]
Towards Comprehensive Description Generation from Factual Attribute-Value Tables
Tianyu Liu, Fuli Luo, Pengcheng Yang, Wei Wu, Baobao Chang, Zhifang Sui
Association for Computational Linguistics (ACL), 2019
[PDF] [bib]
Hierarchical Encoder with Auxiliary Supervision for Neural Table-to-text Generation: Learning Better Representation for Tables
Tianyu Liu, Fuli Luo, Qiaolin Xia, Shuming Ma, Baobao Chang, Zhifang Sui
AAAI Conference on Artificial Intelligence (AAAI), 2019
[PDF] [bib]
Table-to-text Generation by Structure-aware Seq2seq Learning
Tianyu Liu, Kexiang Wang, Lei Sha, Baobao Chang, Zhifang Sui
AAAI Conference on Artificial Intelligence (AAAI), 2018
[PDF] [bib] [code]
A Soft-label Method for Noise-tolerant Distantly Supervised Relation Extraction
Tianyu Liu, Kexiang Wang, Baobao Chang, Zhifang Sui
Empirical Methods in Natural Language Processing (EMNLP), 2017
[PDF] [bib] [code]
Large-scale Simple Question Generation by Template-based Seq2seq Learning
Tianyu Liu, Bingzhen Wei, Baobao Chang, Zhifang Sui
CCF International Conference on Natural Language Processing and Chinese Computing (NLPCC), 2017
[paper] [bib] [code] Best Student Paper
Encoding Temporal Information for Time-Aware Link Prediction
Tingsong Jiang, Tianyu Liu, Tao Ge, Lei Sha, Sujian Li, Baobao Chang, Zhifang Sui
Empirical Methods in Natural Language Processing (EMNLP), 2016
[PDF] [bib]
Towards Time-Aware Knowledge Graph Completion
Tingsong Jiang, Tianyu Liu, Tao Ge, Lei Sha, Baobao Chang, Sujian Li, Zhifang Sui
International Conference on Computational Linguistics (COLING), 2016
[paper] [bib]