Zhenyu Zhu
PhD
Swiss Federal Technology Institute of Lausanne (EPFL)
Mathematical Foundations for Deep Learning

Despite the extraordinary success of deep learning, its emerging weaknesses such as robustness, generalization, and bias, demand an ever closer attention. Unfortunately, many of the existing theories were developed for low capacity models and therefore do not account for the impressive scaling properties of deep learning. For example, enforcing low bias in computer vision tasks using tools developed for linear models have been shown to simply degrade the performance of neural networks. Instead, our work already shows how dedicated theories can shed light on how architectural choices impact robustness and generalization. In this PhD project, we will first investigate how common practices theoretically affect the robustness, generalization, and bias of neural networks. Second, we will propose new grounded approaches improving over existing methods. The overall goal is to improve the reliability of neural networks starting from their theoretical foundations and optimization to improving their behaviour under principled forms of distribution shifts such as covariate and label shifts.

Track:
Industry Track
PhD Duration:
September 1st, 2022 - August 31st, 2026
ELLIS Edge Newsletter
Join the 6,000+ people who get the monthly newsletter filled with the latest news, jobs, events and insights from the ELLIS Network.