Simplified derivations for high-dimensional convex learning problems
David G. Clark, Haim Sompolinsky
SciPost Phys. Lect. Notes 105 (2025) · published 23 October 2025
- doi: 10.21468/SciPostPhysLectNotes.105
- Submissions/Reports
-
Abstract
Statistical-physics calculations in machine learning and theoretical neuroscience often involve lengthy derivations that obscure physical interpretation. Here, we give concise, non-replica derivations of several key results and highlight their underlying similarities. In particular, using a cavity approach, we analyze three high-dimensional learning problems: perceptron classification of points, perceptron classification of manifolds, and kernel ridge regression. These problems share a common structure—a bipartite system of interacting feature and datum variables—enabling a unified analysis. Furthermore, for perceptron-capacity problems, we identify a symmetry that allows derivation of correct capacities through a naïve method.
Authors / Affiliations: mappings to Contributors and Organizations
See all Organizations.- 1 David Clark,
- 2 3 4 Haim Sompolinsky
- 1 Columbia University [CU]
- 2 Harvard University
- 3 האוניברסיטה העברית בירושלים / Hebrew University of Jerusalem [HUJI]
- 4 Harvard Medical School
