DPG Phi
Verhandlungen
Verhandlungen
DPG

Dresden 2026 – scientific programme

Parts | Days | Selection | Search | Updates | Downloads | Help

SOE: Fachverband Physik sozio-ökonomischer Systeme

SOE 17: Focus Session: Physics of AI II (joint session SOE/DY)

SOE 17.3: Talk

Friday, March 13, 2026, 10:15–10:30, GÖR/0226

Online Learning Dynamics and Neural Scaling Laws for a Perceptron Classification Problem — •Yoon Thelge, Marcel Kuhn, and Bernd Rosenow — Institute for Theoretical Physics, University of Leipzig, 04103 Leipzig, Germany

Understanding neural scaling laws and emergence of power law generalisations remains a central challenge in learning dynamics. A natural setting for analysing this behaviour is the online-learning dynamics of a perceptron trained in a teacher*student scenario, where in the thermodynamic limit, the generalisation error exhibits characteristic power-law decay. In realistic classification problems, the teacher is a discrete classifier, while standard gradient-based training requires the student to have continuous outputs. Thus, in practically relevant settings the student is necessarily mismatched to the discrete teacher, a regime that is less well understood. We study this regime for a perceptron with a sign-activation teacher and an error-function student. We derive coupled differential equations for the evolution of the relevant order parameters and verify them via numerical integration and SGD simulations. For fixed learning rates, the generalisation error converges to zero as a power-law with respect to the number of training examples with an exponent of -1/3. The onset of this asymptotic regime shifts with the learning rate, and the generalisation at the onset scales with exponent -1/2, motivating the use of learning-rate schedules to enhance the effective asymptotic decay.

Keywords: Neural Networks; Online learning; Neural scaling laws; Statistical Mechanics

100% | Mobile Layout | Deutsche Version | Contact/Imprint/Privacy
DPG-Physik > DPG-Verhandlungen > 2026 > Dresden