I work at Tencent AI Lab as a research scientist now in Shenzhen.
I graduated from Wuhan University (2017/09 - 2021/06) with a bachelor’s degree and from the Intelligent Interaction Group, SIGS @ Tsinghua University (2021/09 - 2024/06) with a master’s degree, advised by Prof. Yujiu Yang.
Research
I am generally interested in natural language processing and machine learning. Current interests include:
- Effective Model Architecture
- Efficient Training and Inference
- Mathematical Reasoning
Selected Papers
* denotes co-first authors, $^\dagger$ denotes corresponding author/main advisor
DeepMath-103K: A Large-Scale, Challenging, Decontaminated, and Verifiable Mathematical Dataset for Advancing Reasoning
Zhiwei He*, Tian Liang*, Jiahao Xu*, Qiuzhi Liu, Xingyu Chen, Yue Wang, Linfeng Song, Dian Yu, Zhenwen Liang, Wenxuan Wang, Zhuosheng Zhang, Rui Wang$^\dagger$, Zhaopeng Tu$^\dagger$, Haitao Mi, Dong Yu
In submission to Neurips 2025. [arxiv] [code] [bib]
- Top-1 dataset @ HF with 40K+ downloads, Github Stars: 220+
Encouraging Divergent Thinking in Large Language Models through Multi-Agent Debate
Tian Liang*, Zhiwei He*, Wenxiang Jiao*, Xing Wang$^\dagger$, Yan Wang, Rui Wang, Yujiu Yang$^\dagger$, Zhaopeng Tu, Shuming Shi
EMNLP 2024. [arxiv] [code] [bib]
- Pioneer of Multi-Agent LLM, Citations: 520+, Github Stars: 400+
Do NOT Think That Much for 2+3=? On the Overthinking of Long Reasoning Models
Xingyu Chen*, Jiahao Xu*, Tian Liang*, Zhiwei He*, Jianhui Pang, Dian Yu, Linfeng Song, Qiuzhi Liu, Mengfei Zhou, Zhuosheng Zhang, Rui Wang$^\dagger$, Zhaopeng Tu$^\dagger$, Haitao Mi, Dong Yu
ICML 2025. [arxiv] [code] [bib]
- Citations: 170+, Adopted by Kimi-1.5
Critical Tokens Matter: Token-Level Contrastive Estimation Enhances LLM’s Reasoning Capability
Zicheng_Lin*, Tian Liang*, Jiahao Xu*, Qiuzhi Liu, Xing Wang, Ruilin Luo, Chufan Shi, Siheng Li, Yujiu Yang$^\dagger$, Zhaopeng Tu$^\dagger$
ICML 2025. [arxiv] [code] [bib]
Thoughts Are All Over the Place: On the Underthinking of Long Reasoning Models
Yue Wang, Qiuzhi Liu, Jiahao Xu, Tian Liang, Xingyu Chen, Zhiwei He, Linfeng Song, Dian Yu, Juntao Li, Zhuosheng Zhang, Rui Wang, Zhaopeng Tu$^\dagger$, Haitao Mi, Dong Yu
In submission to Neurips 2025. [arxiv] [bib]
Addressing Entity Translation Problem via Translation Difficulty and Context Diversity
Tian Liang, Xing Wang$^\dagger$, Mingming Yang, Yujiu Yang$^\dagger$, Shuming Shi, Zhaopeng Tu
ACL 2024. [paper] [code] [bib]
Exploring Human-Like Translation Strategy with Large Language Models
Zhiwei He*, Tian Liang*, Wenxiang Jiao, Zhuosheng Zhang, Yujiu Yang, Rui Wang$^\dagger$, Zhaopeng Tu$^\dagger$, Shuming Shi, Xing Wang
TACL 2023. [arxiv] [code] [bib]
CriticBench: Benchmarking LLMs for Critique-Correct Reasoning
Zicheng Lin*, Zhibin Gou*, Tian Liang, Ruilin Luo, Haowei Liu, Yujiu Yang$^\dagger$
ACL 2024. [arxiv] [code] [bib]
ParroT: Translating during Chat using Large Language Models tuned with Human Translation and Feedback
Wenxiang Jiao*, Jen-tse Huang, Wenxuan Wang, Zhiwei He, Tian Liang, Xing Wang, Shuming Shi, Zhaopeng Tu
EMNLP 2023. [arxiv] [code] [bib]
Competition
Acronym extraction with hybrid strategies
SDU@ AAAI-22, Rank 2nd.
Siheng Li*, Cheng Yang*, Tian Liang*, Xinyu Zhu, Chengze Yu, Yujiu Yang$^\dagger$
AAAI 2022 Workshop. [paper] [code] [bib]
Multilingual Acronym Disambiguation with Multi-choice Classification
SDU@ AAAI-22, Rank 4th.
Xinyu Zhu*, Chengze Yu*, Siheng Li, Tian Liang, Cheng Yang, Yujiu Yang$^\dagger$
AAAI 2022 Workshop. [paper] [code] [bib]