| 
          Papers
            | 
                Zhiyuan Zeng 曾致远
               
                My name is Zhiyuan Zeng. I am a Ph.D. student (09/2024–present) at the Paul G. Allen School of Computer Science & Engineering, University of Washington, where I have the privilege of being advised by Prof. Hannaneh Hajishirzi and Prof. Pang Wei Koh. I previously received my Bachelor of Engineering degree from the Department of Computer Science and Technology at Tsinghua University in June 2024. During my undergraduate years, I was fortunate to work with Prof. Zhiyuan Liu at Tsinghua University and Prof. Danqi Chen at Princeton University.
               
                Email  / 
                Google Scholar  / 
                Twitter  / 
                Github  / 
                SoP for PhD Application
               |   |  
          * denotes equal contribution
         
           
            | Precise Information Control in Long-Form Text Generation Jacqueline He, Howard Yen*, Margaret Li*, Shuyue Stella Li, Zhiyuan Zeng, Weijia Shi, Yulia Tsvetkov, Danqi Chen, Pang Wei Koh, Luke Zettlemoyer
 NeurIPS 2025
 [paper][code]
 |  
            | Reinforcement Learning for Reasoning in Large Language Models with One Training Example Yiping Wang, Qing Yang, Zhiyuan Zeng, Liliang Ren, Lucas Liu, Baolin Peng, Hao Cheng, Xuehai He, Kuan Wang, Jianfeng Gao, Weizhu Chen, Shuohang Wang, Simon Shaolei Du, Yelong Shen
 NeurIPS 2025
 [paper][code]
 |  
            | EvalTree: Profiling Language Model Weaknesses via Hierarchical Capability Trees Zhiyuan Zeng, Yizhong Wang, Hannaneh Hajishirzi, Pang Wei Koh
 COLM 2025
 [paper][code][demo]
 |  
            | Exploring the Benefit of Activation Sparsity in Pre-training Zhengyan Zhang, Chaojun Xiao, Qiujieli Qin, Yankai Lin, Zhiyuan Zeng, Xu Han, Zhiyuan Liu, Ruobing Xie, Maosong Sun, Jie Zhou
 ICML 2024
 [paper][code]
 |  
            | Evaluating Large Language Models at Evaluating Instruction Following Zhiyuan Zeng, Jiatong Yu, Tianyu Gao, Yu Meng, Tanya Goyal, Danqi Chen
 ICLR 2024
 [paper][code]
 |  
            | Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning Mengzhou Xia, Tianyu Gao, Zhiyuan Zeng, Danqi Chen
 ICLR 2024
 [paper][code][blog]
 |  
            | Emergent Modularity in Pre-trained Transformers Zhengyan Zhang*, Zhiyuan Zeng*, Yankai Lin, Chaojun Xiao, Xiaozhi Wang, Xu Han, Zhiyuan Liu, Ruobing Xie, Maosong Sun, Jie Zhou
 Findings of ACL 2023
 [paper][code]
 |  
            | Plug-and-Play Knowledge Injection for Pre-trained Language Models Zhengyan Zhang*, Zhiyuan Zeng*, Yankai Lin, Huadong Wang, Deming Ye, Chaojun Xiao, Xu Han, Zhiyuan Liu, Peng Li, Maosong Sun, Jie Zhou
 ACL 2023
 [paper][code]
 |  |