Hey there! I'm a first-year PhD student in Computer Science at University of Illinois
Urbana-Champaign
(UIUC). I'm
fortunate to work with Prof.
Yue Guo and Prof. Heng Ji.
I'm currently working on agentic AI,
grounded
multi-modal reasoning, and controllable multi-modal generation.
Previously, I did my Master at University of California, Los Angeles (UCLA),
where I was
advised by Prof. Nanyun
Peng,Prof.
Kai-Wei
Chang and Prof. Yiwei
Wang.
Before that, I got my B.S.in Computer Science and Data Science (double
major) at Purdue University.
My long-term research vision is to build multisensory assistant with adaptive and creative
intelligence that can reason across every modality, and learn new skill/knowledge to adapt to
different context on-the-fly.
Out of curiosity, I have also explored AI for Human Health.
My current research focuses on enhancing agents (e.g. language agent, visual agent, embodied agent,
etc.) with the learning-on-the-job capability: long-horizon run-time learning for
adaptive
multimodal agents that update their knowledge and behavior from in-situ feedback and new context.
My recent works mainly span across two directions: