Skip to content

Latest commit

 

History

History
13 lines (8 loc) · 1.54 KB

methods.md

File metadata and controls

13 lines (8 loc) · 1.54 KB

RF experiments

We validate the RF eigenframework by experimentally probing two random feature models:

  • High-dimensional linear regression with synthetic data generated with powerlaw structure. predictor on synthetic data (Gaussian inputs with labels generated by a powerlaw $\hat y = Wx$
  • Kernel regression with the NNGP kernel of the neural network $f(x)=W_2^T\cdot\mathrm{ReLU}(W_1 x)$ learning CIFAR10. This learning task was studied in (preetums paper).

We emphasize that powerlaw structure is not a requirement for applying our RF eigenframework; we chose synthetic experiments with powerlaw structure because natural data have this structure.

We choose a plot that demonstrates the validity of our theory on both sides of the interpolation threshold. Due to computational constraints this requires us to choose a small trainset size ($n=256$) which results in weak performance even as $k$ increases. With a larger trainset, the test error decreases more substantially as one increases the number of features.

Since both models have the same eigenstructure, we expect them to perform identically in the infinite-feature limit.

We note that it is infeasible to accurately measure the asymptotic powerlaw exponents by directly fitting the finite-size spectra. Thankfully, there are techniques to measure the exponents by proxy. These techniques (discussed in TODO) yield precise results and we employ them throughout our experiments. Exact powerlaws with the measured exponents are shown here in dotted lines (vertically shifted for visual clarity).