Headlines
-
A scenario for learning in overparametrized neural networks
Modern neural networks, with billions of parameters, are so overparametrized that they can “overfit” even random, structureless data. Yet when trained on datasets with structure, they learn the underlying features. Understanding why overparametrization does not destroy their effectiveness is a fundamental challenge in AI. Two researchers, Andra Montanari (Stanford) and Pierfrancesco Urbani (IPhT) propose that…

Agenda
15 December 2025
11h00 – 12h30
From cosmology to statistical physics and condensed matter: Burgers and Gross-Pitaevskii equations
Salle Claude Itzykson, Bât. 774
16 December 2025
14h00 – 15h30
TBA
Salle Claude Itzykson, Bât. 774
5 January 2026
11h00 – 12h30
La FK-percolation presque critique en milieu aléatoire
Salle Claude Itzykson, Bât. 774
Aucun événement












