Tesi etd-05022022-184611 |
Link copiato negli appunti
Tipo di tesi
Tesi di laurea magistrale
Autore
BASTERI, ANDREA
URN
etd-05022022-184611
Titolo
Quantitative Convergence of Randomly Initialized Wide Deep Neural Networks Towards Gaussian Processes
Dipartimento
MATEMATICA
Corso di studi
MATEMATICA
Relatori
relatore Trevisan, Dario
Parole chiave
- ANN
- deep learning
- deep neural networks
- DNN
- Gaussian Processes
- machine learning
- neural networks
- numerical experiments
- optimal transport
- Wasserstein distance
Data inizio appello
10/06/2022
Consultabilità
Completa
Riassunto
The thesis regards the quantitative convergence of randomly initialized fully connected deep neural networks towards a suitable Gaussian limit. In particular we estimate the 2-Wasserstein distance between the joint distribution of the outputs of a neural network and a suitable Gaussian Process, the "Neural Network Gaussian Process" (NNGP).
We make our estimates with techniques of optimal transport and techniques linked to Gaussian processes, and we obtain that such rate of convergence can be estimated with the inverse of the square root of the numbers of neurons in the layers.
We provide some numerical simulation to show that our estimates are not sharp, so further studies are needed to find the true rate of convergence and to study other properties of this convergence.
We make our estimates with techniques of optimal transport and techniques linked to Gaussian processes, and we obtain that such rate of convergence can be estimated with the inverse of the square root of the numbers of neurons in the layers.
We provide some numerical simulation to show that our estimates are not sharp, so further studies are needed to find the true rate of convergence and to study other properties of this convergence.
File
Nome file | Dimensione |
---|---|
Master_T...itiva.pdf | 6.09 Mb |
Contatta l’autore |