21–26 Jun 2026
U. Ottawa - Learning Crossroads (CRX) Building
America/Toronto timezone
Welcome to the 2026 CAP Congress Program website! / Bienvenue au siteweb du programme du Congrès de l'ACP 2026!

Simmering: Physical is Better than Optimal for Training Neural Networks

Not scheduled
15m
U. Ottawa - Learning Crossroads (CRX) Building

U. Ottawa - Learning Crossroads (CRX) Building

100 Louis-Pasteur Private, Ottawa, ON K1N 9N3
Oral Competition (Graduate Student) / Compétition orale (Étudiant(e) du 2e ou 3e cycle) Condensed Matter and Materials Physics / Physique de la matière condensée et matériaux (DCMMP-DPMCM) (DCMMP) T1-4 | (DPMCM)

Speaker

Irina Babayan

Description

Neural networks model complex phenomena within and beyond physics by leveraging their universal estimation capacity. However, this universality interferes with generalizability when combined with optimization-based training. This tension is exemplified by the prevalence of overfitting, where naive optimization drives networks into overspecified representations that deviate from underlying phenomena. Phenomena can be better represented by collections of networks, but many so-called “ensemble”-based approaches are still undermined by the intrinsic difficulty of reconciling optimization-based training with noisy data. Nonetheless, observed improvements from “ensemble”-based methods suggest that less-haphazard ensemble generation approaches that leverage physical insight could improve neural network training.

In this talk, we show that physics-based ensemble methods outperform optimization-based training. Rather than seeking optimal networks, our sufficient-training approach generates “good enough” weights and biases, which, paradoxically, outperform leading optimization-based approaches. We resolve this apparent paradox using information theory to show that sufficient training provides a minimally-biased representation of the underlying phenomenon. Our results show that straightforward, physics-based approaches can supplant optimization for training transformers, and feedforward and convolutional neural networks, by offering superior training performance and powerful insights for neural networks in physics applications and beyond.

Keyword-1 Neural networks
Keyword-2 Statistical Physics

Author

Irina Babayan

Co-authors

Hazhir Aliahmadi (Queens University) Greg van Anders (Queen's University)

Presentation materials

There are no materials yet.