SciPost Submission Page
Random features and polynomial rules
by Fabián Aguirre-López, Silvio Franz, Mauro Pastore
Submission summary
| Authors (as registered SciPost users): | Mauro Pastore |
| Submission information | |
|---|---|
| Preprint Link: | scipost_202407_00018v2 (pdf) |
| Code repository: | https://github.com/MauroPastore/RandomFeatures/ |
| Date accepted: | Dec. 10, 2024 |
| Date submitted: | Nov. 7, 2024, 4:49 p.m. |
| Submitted by: | Mauro Pastore |
| Submitted to: | SciPost Physics |
| Ontological classification | |
|---|---|
| Academic field: | Physics |
| Specialties: |
|
| Approaches: | Theoretical, Computational |
Abstract
Random features models play a distinguished role in the theory of deep learning, describing the behavior of neural networks close to their infinite-width limit. In this work, we present a thorough analysis of the generalization performance of random features models for generic supervised learning problems with Gaussian data. Our approach, built with tools from the statistical mechanics of disordered systems, maps the random features model to an equivalent polynomial model, and allows us to plot average generalization curves as functions of the two main control parameters of the problem: the number of random features N and the size P of the training set, both assumed to scale as powers in the input dimension D. Our results extend the case of proportional scaling between N, P and D. They are in accordance with rigorous bounds known for certain particular learning tasks and are in quantitative agreement with numerical experiments performed over many order of magnitudes of N and P. We find good agreement also far from the asymptotic limits where D → ∞ and at least one between P/D^K , N/D^L remains finite.
Author indications on fulfilling journal expectations
- Provide a novel and synergetic link between different research areas.
- Open a new pathway in an existing or a new research direction, with clear potential for multi-pronged follow-up work
- Detail a groundbreaking theoretical/experimental/computational discovery
- Present a breakthrough on a previously-identified and long-standing research stumbling block
Published as SciPost Phys. 18, 039 (2025)
Reports on this Submission
Report
Recommendation
Publish (meets expectations and criteria for this Journal)
Report #1 by Anonymous (Referee 1) on 2024-11-13 (Invited Report)
- Cite as: Anonymous, Report on arXiv:scipost_202407_00018v2, delivered 2024-11-13, doi: 10.21468/SciPost.Report.10110
Strengths
I am re-listing all the strengths of this work that I have already underlined in my report for the first version of this paper:
- The authors provide a tight asymptotic characterization of the learning of Random Features Models (RFM) on a random polynomial target function, in various data/width/dimension regimes.
-They outline and identify data (resp. width) limited regimes where the RFM reduces to a kernel (resp. polynomial regression) method, as well as a non-trivial width~data regime, exhibiting in particular an interpolation peak phenomenon.
-The derivation relies on the replica method from statistical physics and several random matrix theory arguments and approximations. All steps are rather clearly justified, motivated, and discussed.
-
The analytical findings are supported by convincing numerics.
-
The consequences/takeaways of the analytical results are discussed, notably in terms of overfitting and expressive power.
Weaknesses
Report
Recommendation
Publish (meets expectations and criteria for this Journal)
