I study how agents can learn to act in complex, multi-agent environments with minimal experience by exploiting structure in the underlying decision problem. My interests are at the intersection of reinforcement learning, control theory, and multi-agent systems. I focus on sample complexity reduction through latent structure in MDPs, including low-rank Q-functions, spectral gaps, and manifold geometry.
Matrix Completion
Low-rank Q-function recovery from sparse samples via leveraged CUR decompositions.
Subspace Recovery
Two-to-infinity error control for row-wise accuracy in policy extraction.
World Models
Score matching in learned latent spaces for planning guarantees scaling with intrinsic dimension.
Imitation Learning
Reducing demonstration complexity when the expert's Q-function has low-rank structure.
| Arbabian Lab | Stanford University |
| EMERGE Lab | New York University |
| MVA Program | ENS Paris-Saclay |
| Mathematics | Sorbonne University |