I am an individual researcher working on large language models, multimodal systems, and agentic learning.
I care about a simple question: how do we move from models that generate convincing outputs to systems that can reason, act, and improve through interaction with the world? That question sits behind my interest in long-horizon planning, tool use, multimodal representation, and training methods that make model behavior more grounded.
My recent work spans agentic coding, multimodal modeling, reinforcement-learning methods, and efficient attention design. Across these projects, I keep returning to the same underlying theme: useful intelligence depends not only on model scale, but also on how architectures, environments, and training pipelines are designed together.
If you want a quick overview, start with the selected papers below or the essays in the blog.
Latest essay · March 26, 2026
How the AI field is transitioning from static reasoning models to agentic systems that think in order to act.
Read the essay