```html
Student weights \(W\) learn to match a fixed teacher on Gaussian inputs using AdamW, with added gradient noise.
Outputs are normalized: \(y = \frac{Wx}{\|Wx\|}\).
Solid lines = empirical; dashed lines = predicted equilibrium values using the more precise
\(W_\infty\) formula from the blog (no dependence on \(W_0\)). Moving sliders re-bases the dashed lines.