Back-to-School Seminar -- Spline time: Using geometry to understand deep neural networks

Grace O'Brien, University of Washington
-
PDL C-38

Have you heard the words "neural networks" but not totally understood what they are? Are you interested in hearing about my experience interning at PNNL? Do you like pretty pictures? Come to my talk and get an introduction to machine learning and the math behind it. I'll also discuss my specific project this summer, described below.

AI systems are increasingly becoming a part of our everyday lives, including in safety-critical systems. However, the innerworkings of deep neural networks are still largely a black box. Even in the case of classification tasks, common methods used to assess model performance do not give insight into whether the model will generalize or into other performance nuances. In this project, we use mathematical techniques to better understand how these processes work and explore how to identify problems such as overfitting, memorization, and poor generalization.  In the process of training a model, piecewise-linear activation functions partition the input space into two, creating a tiling that shifts over time. Following the work of Balestriero, Baraniuk, and others, we use geometric tools to study this tiling to gain insight into the model’s training progress and, potentially provide greater assurances that a model is ready for deployment.
 

Event Type
Event Subcalendar