Yinghui He 何映晖
Hi there! I’m Yinghui He (pronounced as Yee·ng-Hway Huh), a second-year PhD student at Princeton University Computer Science Department. I’m honored to be advised by Sanjeev Arora at Princeton Language and Intelligence (PLI). I work in Natural Language Processing, pursuing to develop LLMs and agents to understand the two-way relation between artificial intelligence and human cognition. I finished my Bachelor’s degree in Computer Science at the University of Michigan, where I had the honor to work with Rada Mihalcea (at the LIT Lab) and Wei Hu.
Publications
AdaptMI: Adaptive Skill-based In-context Math Instruction for Small Language Models
Yinghui He, Abhishek Panigrahi, Yong Lin, Sanjeev Arora
COLM 2025; ICML 2025 Workshop on Test-Time Adaptation; ICML 2025 Methods and Opportunities at Small Scale Workshop
EmoAgent: Assessing and Safeguarding Human-AI Interaction for Mental Health Safety
Jiahao Qiu*, Yinghui He*, Xinzhe Juan*, Yiming Wang, Yuhan Liu, Zixin Yao, Yue Wu, Xun Jiang, Ling Yang, Mengdi Wang
EMNLP 2025 Main Conference
LongProc: Benchmarking Long-Context Language Models on Long Procedural Generation
Xi Ye, Fangcong Yin*, Yinghui He*, Joie Zhang*, Howard Yen*, Tianyu Gao, Greg Durrett, Danqi Chen
COLM 2025
Hi-ToM: A Benchmark for Evaluating Higher-Order Theory of Mind Reasoning in Large Language Models
Yinghui He, Yufan Wu, Yilin Jia, Rada Mihalcea, Yulong Chen, and Naihao Deng
Findings of EMNLP 2023; ICML 2023 Workshop on Theory of Mind in Communicating Agents
Robust Sparse Mean Estimation via Incremental Learning
Jianhao Ma, Rui Ray Chen, Yinghui He, Salar Fattahi, and Wei Hu
ICLR 2024 Workshop on Bridging the Gap Between Practice and Theory in Deep Learning