logo SBA

ETD

Archivio digitale delle tesi discusse presso l’Università di Pisa

Tesi etd-02132025-203730


Tipo di tesi
Tesi di laurea magistrale
Autore
CARDIA, ANDREA
URN
etd-02132025-203730
Titolo
Minimizing Entropy for Training and Quantization (METaQ): A Novel Algorithm for Neural Network Training and Compression
Dipartimento
INFORMATICA
Corso di studi
DATA SCIENCE AND BUSINESS INFORMATICS
Relatori
relatore Prof. Frangioni, Antonio
correlatore Prof. Ferragina, Paolo
Parole chiave
  • artificial intelligence
  • bundle methods
  • compression
  • neural network
  • neural network compression
  • non-differentiable problem
  • optimization
  • quantization
  • subgradient methods
Data inizio appello
28/02/2025
Consultabilità
Completa
Riassunto
A neural network compression strategy is developed during the training phase by adding a regularization term $\phi(w)$ to the loss function, in order to minimize the entropy of the network weights. $\phi(w)$ results from a nondifferentiable optimization problem that is computationally very complex to solve. However, fortunately, to train the network, it is not necessary to explicitly find $\phi(w)$; instead, it is sufficient to provide its (sub)gradient to standard machine learning tools (such as PyTorch) to guide the training towards a low entropy weight configuration (in addition to achieving good accuracy). This subgradient can be computed using the optimal Lagrange multipliers $\beta^*$ associated with the set of constraints involving the weights $w$ in the problem that defines $\phi(w)$.

In this work, we will develop a procedure to determine $\beta^*$ by applying Lagrangian relaxation and optimization techniques, also developing ad hoc methods for certain sub-problems when necessary for efficiency reasons. Once a trained network with low entropy is obtained, the compression strategy culminates in the quantization of the weights. The task of encoding the weights is implemented via well-known compression algorithms that come arbitrarily close to the entropy.

The introduction of the term $\phi(w)$ appropriately modeled for the function to be optimized during training is what makes this work innovative.

The tests were conducted using the LeNet-5 network on the MNIST dataset, although the strategy is also applicable to larger networks. The achieved results show a $29\times$ compression of LeNet-5 (Compression Ratio 3.43\%) with an accuracy of 99.01\%, making METaQ a strategy comparable to state-of-the-art NN-compression algorithms.

The project's source code is freely available at: https://github.com/cardiaa/METaQ
File