I study how agents can learn to act in complex, multi-agent environments with minimal experience by exploiting structure in the underlying decision problem. My interests are at the intersection of reinforcement learning, control theory, and multi-agent systems. I focus on sample complexity reduction through latent structure in MDPs, including low-rank Q-functions, spectral gaps, and manifold geometry.

Matrix Completion

Low-rank Q-function recovery from sparse samples via leveraged CUR decompositions.

Subspace Recovery

Two-to-infinity error control for row-wise accuracy in policy extraction.

World Models

Score matching in learned latent spaces for planning guarantees scaling with intrinsic dimension.

Imitation Learning

Reducing demonstration complexity when the expert's Q-function has low-rank structure.

Arbabian LabStanford University
EMERGE LabNew York University
MVA ProgramENS Paris-Saclay
MathematicsSorbonne University
Sep 2025 Started MVA at ENS Paris-Saclay