Abstract: We introduce a new framework for the theoretical analysis of learning algorithms based on the new notion of self-regularization, meaning that the algorithm itself produces sufficiently regular functions. As a central example, we analyze gradient descent and show that this framework yields minmax-optimal learning rates in broad settings with comparatively little technical effort.
21
Apr
MIP Seminar: Max Schölpple (University of Stuttgart)
Termin:
- Di.:
- 16:15 - 18:00 Uhr
21. April 2026