2025
DeepTheorem: Advancing LLM Reasoning for Theorem Proving Through Natural Language and Reinforcement Learning.
CoRR, May, 2025

Two Experts Are All You Need for Steering Thinking: Reinforcing Cognitive Effort in MoE Reasoning Models Without Additional Training.
CoRR, May, 2025

DeepMath-103K: A Large-Scale, Challenging, Decontaminated, and Verifiable Mathematical Dataset for Advancing Reasoning.
CoRR, April, 2025

Dancing with Critiques: Enhancing LLM Reasoning with Stepwise Natural Language Self-Critique.
CoRR, March, 2025

The First Few Tokens Are All You Need: An Efficient and Effective Unsupervised Prefix Fine-Tuning Method for Reasoning Models.
CoRR, March, 2025

Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs.
CoRR, January, 2025

Insight Over Sight: Exploring the Vision-Knowledge Conflicts in Multimodal LLMs.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

2024
Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs.
CoRR, 2024

Chain-of-Jailbreak Attack for Image Generation Models via Editing Step by Step.
CoRR, 2024

2023
Simple and Scalable Nearest Neighbor Machine Translation.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

2016
History question classification and representation for Chinese Gaokao.
Proceedings of the 2016 International Conference on Asian Language Processing, 2016