We validate the RF eigenframework by experimentally probing two random feature models:
- High-dimensional linear regression with synthetic data generated with powerlaw structure. predictor on synthetic data (Gaussian inputs with labels generated by a powerlaw
$\hat y = Wx$ - Kernel regression with the NNGP kernel of the neural network
$f(x)=W_2^T\cdot\mathrm{ReLU}(W_1 x)$ learning CIFAR10. This learning task was studied in (preetums paper).
We emphasize that powerlaw structure is not a requirement for applying our RF eigenframework; we chose synthetic experiments with powerlaw structure because natural data have this structure.
We choose a plot that demonstrates the validity of our theory on both sides of the interpolation threshold. Due to computational constraints this requires us to choose a small trainset size (
Since both models have the same eigenstructure, we expect them to perform identically in the infinite-feature limit.
We note that it is infeasible to accurately measure the asymptotic powerlaw exponents by directly fitting the finite-size spectra. Thankfully, there are techniques to measure the exponents by proxy. These techniques (discussed in TODO) yield precise results and we employ them throughout our experiments. Exact powerlaws with the measured exponents are shown here in dotted lines (vertically shifted for visual clarity).