About me

I am a forth-year PhD student (2021 - present) of Department of Computer Science and Engineering at Shanghai Jiao Tong University (SJTU). I am fortunate to be advised by Prof. Rui Wang. Before that, I received the bachelor degree in Software Engineering from South China University of Technology (SCUT). I am currently a research intern at Tencent AI Lab, co-advised by Dr. Xing Wang and Dr. Zhaopeng Tu. I also work closely with Zhuosheng Zhang.

🔬 Research

Large and Efficient Reasoning Models

  • Underthinking issue in o1-like models [Preprint]
  • Overthinking issue in o1-like models [Preprint]
  • Rank-sharing LoRA [ICLR 2025]

Autonomous Agent powered by Large Language Models

  • Multi-agent debate [EMNLP 2024]
  • Evaluating and improving agent safety [EMNLP 2024 (Findings)]

Multilinguality & Machine Translation

  • Bridging the gap between training signal and real user input [ACL 2022]
  • Human-like translation strategy [TACL 2024]
  • Improving translation with human feedback [NAACL 2024]
  • Cross-lingual consistency for text watermark [ACL 2024 (Oral)]

🔥 News

  • 2025.01: 🤯🤯 Revealed underthinking issue in o1-like models (preprint).
  • 2024.12: 🎉🎉 One paper about parameter-efficient fine-tuning accepted by ICLR 2025.
  • 2024.12: 🤯🤯 Revealed overthinking issue in o1-like models (preprint).
  • 2024.08: 🇹🇭🐘 Gave an oral presentation at ACL 2024 on cross-lingual text watermark.
  • 2024.06: 🇲🇽🌮 Attended NAACL 2024 @ Mexico.
  • 2024.05: 🎉🎉 One paper about cross-lingual text watermark accepted by ACL 2024.
  • 2024.03: 🎉🎉 One paper about improving translation with human feedback accepted by NAACL 2024.
  • 2023.11: 🎉🎉 One paper about human-like translation strategy accepted by TACL 2024.
  • 2023.05: Introduced the MAPS framework, enabling LLMs to mimic the human translation strategy. See also the media coverage 📸.
  • 2023.05: Proposed a multi-agent debate framework (MAD) with large language models (EMNLP 2024).

🖨️ Selected preprints

* denotes co-first authors

arXiv 2025
wang2025thought

Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs

Yue Wang*, Qiuzhi Liu*, Jiahao Xu*, Tian Liang*, Xingyu Chen*, Zhiwei He*, Linfeng Song, Dian Yu, Juntao Li, Zhuosheng Zhang, Rui Wang, Zhaopeng Tu, Haitao Mi, Dong Yu

o1-like models underthink, which:

  1. Occur more frequently on harder problems,
  2. Lead to frequent switching between thoughts without reaching a conclusion,
  3. Correlate with incorrect responses due to insufficient exploration.
arXiv 2024
chen2024not

Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs

Xingyu Chen*, Jiahao Xu*, Tian Liang*, Zhiwei He*, Jianhui Pang, Dian Yu, Linfeng Song, Qiuzhi Liu, Mengfei Zhou, Zhuosheng Zhang, Rui Wang, Zhaopeng Tu, Haitao Mi, Dong Yu

o1-like models overthink, which:

  1. Contribute minimally to accuracy.
  2. Lack diversity in reasoning strategies.
  3. Occur more frequently with simple problems.

📝 Selected publications

* denotes co-first authors

🎖 Honors and Awards

  • 2022.8: 1st place in the WMT22 General Translation Task, English to Livonian (Unconstrained System).
  • 2022.8: 2nd place in the WMT22 General Translation Task, Livonian to English (Unconstrained System).
  • 2018, 2019: First Class Scholarship.

💬 Invited Talks

  • 2024.07: Can Watermarks Survive Translation, AITIME | [video] [slide]
  • 2023.11: Improving Machine Translation with Human Strategy and Feedback, CJNLP | [slide]
  • 2022.08: Unsupervised Neural Machine Translation, CCKS 2022

💻 Internships