Zhenting Qi

漆振霆 | [ch'i chen t'ing]

me.jpg

Welcome! I am an incoming Computer Science Ph.D. student at Harvard University, where I am honored to be co-advised by Prof. Yilun Du and Prof. Hima Lakkaraju. My research focuses on advancing Foundation Models and Generative AI, with the ultimate vision of developing intelligent and reliable AI systems that benefit human society. Motivated by this, I am generally interested in the following topics (w/o particular order):

  • Reasoning
    • Understanding and enhancing reasoning capabilities in foundation models
    • Developing AI systems that generalize effectively to OOD scenarios
    • Training (multi-)agents for compositional reasoning tasks
  • Reliability
    • Improving understanding of foundation models and AI systems
    • Enhancing AI controllability and robustness
    • Designing scalable methods to ensure reliability while advancing capabilities

More specifically, I am currently investigating several exciting directions:

  • 🤖 Training agents for coding assistance and scientific discovery
  • 🧠 Developing advanced memory mechanisms for agents
  • 🔄 Training-time and test-time self-evolution
  • 📊 Dynamic evaluation for reasoning/generalization of foundation models and agents

Our research group actively welcomes collaborations, and I am always excited to chat about research ideas! Please feel free to reach out at: zhentingqi [at] g [dot] harvard [dot] edu

For more information about my research, please see Google Scholar, Semantic Scholar, or DBLP.

About Me

I hold a master’s degree in Computational Science and Engineering from Harvard, and dual bachelor’s degrees in Computer Engineering from UIUC and ZJU (highest honors). I am also a recipient of Harvard SEAS Prize Fellowship.

I’ve had the privilege of working closely with many distinguished researchers, including (the late) Prof. Dragomir R. Radev at Yale, Prof. Volodymyr Kindratenko at UIUC, Dr. Li Lyna Zhang at Microsoft Research Asia, Prof. Chuang Gan at MIT-IBM Watson AI Lab, and Prof. James Glass at MIT.

News

May 30, 2025

Our paper Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search 【新智元】 has been accepted to ICML 2025.

May 01, 2025

Will be joining Google DeepMind (Mountain View office) as a Student Researcher!

Apr 15, 2025

I will continue my research journey at Harvard as a PhD student!

Jan 30, 2025

Our papers Mutual Reasoning Makes Smaller LLMs Stronger Problem-Solvers 【机器之心】, Quantifying Generalization Complexity for Large Language Models, Follow My Instruction and Spill the Beans: Scalable Data Extraction from Retrieval-Augmented Generation Systems have been accepted to ICLR 2025.

Selected Publications

  1. satori.png
    Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search
    Maohao Shen*, Guangtao Zeng*, Zhenting Qi*, Zhang-Wei Hong, Zhenfang Chen, and 5 more authors
    In the Forty-Second International Conference on Machine Learning (ICML), 2025
  2. rstar.png
    Mutual Reasoning Makes Smaller LLMs Stronger Problem-Solver
    Zhenting Qi, Mingyuan MA, Jiahang Xu, Li Lyna Zhang, Fan Yang, and 1 more author
    In The Thirteenth International Conference on Learning Representations (ICLR), 2025
  3. scylla.png
    Quantifying Generalization Complexity for Large Language Models
    Zhenting Qi, Hongyin Luo, Xuliang Huang, Zhuokai Zhao, Yibo Jiang, and 3 more authors
    In The Thirteenth International Conference on Learning Representations (ICLR), 2025
  4. rag.png
    Follow My Instruction and Spill the Beans: Scalable Data Extraction from Retrieval-Augmented Generation Systems
    Zhenting Qi, Hanlin Zhang, Eric P. Xing, Sham M. Kakade, and Himabindu Lakkaraju
    In The Thirteenth International Conference on Learning Representations (ICLR), 2025
  5. loft.png
    LoFT: Enhancing Faithfulness and Diversity for Table-to-Text Generation via Logic Form Control
    Yilun Zhao*, Zhenting Qi*, Linyong Nan, Lorenzo Jaime Flores, and Dragomir Radev
    In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, May 2023
  6. pillow.png
    PILLOW: Enhancing Efficient Instruction Fine-tuning via Prompt Matching
    Zhenting Qi, Xiaoyu Tan, Shaojie Shi, Chao Qu, Yinghui Xu, and 1 more author
    In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track, Dec 2023
  7. safer.png
    SaFER: A Robust and Efficient Framework for Fine-tuning BERT-based Classifier with Noisy Labels
    Zhenting Qi, Xiaoyu Tan, Chao Qu, Yinghui Xu, and Yuan Qi
    In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 5: Industry Track), Jul 2023
  8. folio.png
    FOLIO: Natural Language Reasoning with First-Order Logic
    Simeng Han, Hailey Schoelkopf, Yilun Zhao, Zhenting Qi, Martin Riddell, and 30 more authors
    In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP), Nov 2024