![School of Natural Sciences Event](/sites/default/files/styles/two_column_medium/public/2019-09/sns_default.jpg?itok=IEu1CLXj)
Princeton University PCTS Workshop
Physics for Neural Networks
Registration is required on line at the PCTS website https://pcts.princeton.edu/events/2023/physics-neural-networks
https://forms.gle/EZ4A7n8MrBDFoJex6
Details can be found on the website at Princeton Center for Theoretical Science
Please Note: This workshop is not open to the general public, but only to active researchers.
Organizers: William Bialek; Giorgio Cipolloni; Boris Hanin; Francesca Mignacco
A hallmark of modem machine learning, exemplified by deep learning, is the use of exceptionally large models to fit complex datasets. While the empirical success of the neural network models used in deep learning is undeniable, a predictive formalism for describing and improving them is nascent. On the theoretical side a promising avenue for making progress is to view neural networks as large collections of (often weakly) interacting parameters. From this perspective, it is natural to study them using the tools of modern physics coming from random matrix theory, statistical mechanics, the 1/n expansion, field theory, and so on. On the experimental side, neural networks present a unique opportunity for impactful empirical investigation in which the microscopic ground truth is known to arbitrary precision. The purpose of this workshop is to bring together a range of scientists trained in a variety of disciplines related to modern physics that are interested in developing both the theory and phenomenology of deep learning.