Traces of Class/CrossClass Structure Pervade Deep Learning Spectra
Numerous researchers recently applied empirical spectral analysis to the study of modern deep learning classifiers, observing spectral outliers and small but distinct bumps often seen beyond the edge of a “main bulk”. This talk presents an important formal class/crossclass structure and shows how it lies at the origin of these visually striking patterns. The structure is shown to permeate the spectra of deepnet features, backpropagated errors, gradients, weights, Fisher Information matrix and Hessian, whether these are considered in the context of an individual layer or the concatenation of them all. The significance of the structure is illustrated by (i) proposing a correction to KFAC, a well known secondorder optimization algorithm for training deepnets, and (ii) proving in the context of multinomial logistic regression that the ratio of outliers to bulk in the spectrum of the Fisher information matrix is predictive of misclassification.
References

The Full Spectrum of Deepnet Hessians at Scale: Dynamics with SGD Training and Sample Size, Vardan Papyan, 2018.

Measurements of ThreeLevel Hierarchical Structure in the Outliers in the Spectrum of Deepnet Hessians, Vardan Papyan, ICML 2019.
Vardan Papyan received his B.Sc. degree in 2013 from the Computer Science Department at the Technion  Israel Institute of Technology, Haifa, Israel. He later received his M.Sc. and Ph.D. in 2017 from the same department, under the supervision of Prof. Michael Elad. He is currently a postdoctoral researcher, advised by Prof. David Donoho, at the Department of Statistics at Stanford University. His research interests include machine learning and more specifically deep learning, data science, signal and image processing, and highdimensional statistics.