Tesi etd-10092021-004215 |
Link copiato negli appunti
Tipo di tesi
Tesi di laurea magistrale
Autore
BUTTAZZO, SERGIO
URN
etd-10092021-004215
Titolo
Asymptotic Dynamics of Adaptive Stochastic Gradient Algorithms
Dipartimento
MATEMATICA
Corso di studi
MATEMATICA
Relatori
relatore Prof. Romito, Marco
Parole chiave
- adadelta
- adagrad
- adam
- adaptive
- algorithm
- asymptotic
- batch
- convergence
- deep learning
- descent
- gradient
- learning rate
- machine learning
- momentum
- optimization
- stochastic
Data inizio appello
29/10/2021
Consultabilità
Completa
Riassunto
Gradient-based optimization algorithms, in particular their stochastic counterparts, have become by far the most widely used algorithms in a lot of applied mathematics problems, first of all statistical modeling and machine learning. After a quick introduction to gradient descent and its stochastic counterpart, the focus is put on variance reduction techniques such as learning rate reduction, batch size increase and momentum. Results about optimal scheduling for these techniques are given in the case of convex loss functions. In the non-convex case, a description of adaptive descent algorithm is given, as well as some theoretical results about their convergence. The framework of stochastic modified equations, in which descent algorithms can be interpreted as discretizations of stochastic differential equations gets introduced and ultimately extended to include adaptive algorithms.
File
Nome file | Dimensione |
---|---|
Tesi_Magistrale.pdf | 986.04 Kb |
Contatta l’autore |