
Sign up to save your podcasts
Or
This work examines the critical points of random neural networks, particularly as network depth increases in the infinite-width limit. The authors provide asymptotic formulas for the expected number of critical points, categorized by their index or when exceeding a threshold. Their analysis reveals three distinct scaling regimes for the expected number of critical points based on a specific derivative of the covariance function. Theoretical findings are corroborated by numerical experiments, which also suggest potential divergence in critical points for networks with irregular activation functions like ReLU.
This work examines the critical points of random neural networks, particularly as network depth increases in the infinite-width limit. The authors provide asymptotic formulas for the expected number of critical points, categorized by their index or when exceeding a threshold. Their analysis reveals three distinct scaling regimes for the expected number of critical points based on a specific derivative of the covariance function. Theoretical findings are corroborated by numerical experiments, which also suggest potential divergence in critical points for networks with irregular activation functions like ReLU.