Hey! I'm a first-year PhD student in Computer Science at University of Illinois Urbana-Champaign
(UIUC). I'm
fortunate to work with Prof.
Yue Guo and Prof. Heng Ji.
I'm currently working on agentic AI,
grounded
multi-modal reasoning, and controllable multi-modal generation.
Previously, I did my Master's Degree at University of California, Los Angeles (UCLA),
where I was
advised by Prof. Nanyun
Peng, Prof. Kai-Wei Chang, and Prof. Yiwei Wang.
Before that, I completed my Bachelor of Science degree in Computer Science and Data Science (double
major) at Purdue University.
My long-term research vision is to build multisensory assistant with adaptive and creative
intelligence that can reason across every modality and learn new skill/knowledge to adapt to
different context on-the-fly.
Out of curiosity, I have also explored AI for Human Health.
My current research focuses on enabling agents (e.g. language agent, visual agent, embodied agent,
etc.) to learn on the job: long-horizon run-time learning for adaptive
multimodal agents that update their knowledge and behavior from in-situ feedback and new context.
My recent works mainly span across two directions: