Complexity of learning
When, and why, can we efficiently learn in high dimensions? What structural properties of data make learning computationally feasible?
Modern data is often high-dimensional: language models work over massive vocabularies, and biological systems involve thousands of interacting components. Yet learning in these settings can be surprisingly effective. A possible explanation is the presence of latent low-dimensional structure. Although the ambient dimension is large, the true complexity of the data is often governed by hidden structure that lives in an essentially much smaller space. Exploiting this structure is key to efficient learning.
I am interested in understanding when such models can be learned efficiently. A recurring theme is the statistical–computational tradeoff: some problems are learnable in principle, but no efficient algorithm achieves the optimal rates. Frameworks such as statistical query (SQ) and low-degree polynomial (LDP) methods help formalize these limitations, while algorithmic work aims to match (or nearly match) these bounds.