I am a 2nd year MSR student at Carnegie Mellon University, where I am fortunately advised by Yonatan Bisk.
Given fundamental limitations in LLM compositionality, sample efficiency, and out-of-distribution generalization, I have pivoted to working on bridging language modeling with structured probabilistic representations for more robust, generalizable world modeling. I'm most interested in integrating language models with program induction to build coherent, theory-based world models that learn new skills from a few examples and update themselves through minimal, local repairs without catastrophic forgetting.
I was first introduced to research at MIT's Computational Reactor Physics Group during my freshman year at Harvard, and began working with machine learning at Caltech's Autonomous Robotics and Controls Lab under Soon-Jo Chung.
University: first last @ andrew.cmu.edu
Personal: first last + ML @ gmail.com