ETD

Archivio digitale delle tesi discusse presso l'Università di Pisa

Tesi etd-05302019-162625


Tipo di tesi
Tesi di laurea magistrale
Autore
MELONI, ENRICO
Indirizzo email
enrico.meloni@outlook.it
URN
etd-05302019-162625
Titolo
Using Virtual Worlds to Train an Object Detector for Personal Protection Equipment
Dipartimento
INGEGNERIA DELL'INFORMAZIONE
Corso di studi
COMPUTER ENGINEERING
Relatori
relatore Prof. Amato, Giuseppe
correlatore Prof. Falchi, Fabrizio
correlatore Gennaro, Claudio
correlatore Dott. Di Benedetto, Marco
Parole chiave
  • safety equipment detection
  • personal protection equipment
  • object detection
  • domain adaption
  • deep learning
  • transfer learning
  • virtual dataset
Data inizio appello
21/06/2019
Consultabilità
Completa
Riassunto
Neural Networks are an effective technique in the field of Artificial Intelligence and in the field of Computer Vision. They learn from examples, without programming into them any previous knowledge. Deep Neural Networks saw many successful applications, thanks to the huge amount of data that is available with the growth of the internet.
When annotations are not available, images are manually annotated introducing very high costs. In some contexts, gathering valuable images could be impractical for reasons related to privacy, copyright and security. To overcome these limitations the research community has taken interest in creating virtual worlds for the generation of automatically annotated training samples. In previous works, using a graphics engine for augmenting a training is shown to be a valid solution.
In this work, we applied the virtual environment to approach to a not yet considered task: the detection of personal protection equipment. We developed V-DAENY, a plugin for GTA-V. With it, we generated over 140,000 automatically annotated images in several locations with different weather conditions. We manually annotated two real datasets for testing. We trained a network with this approach and evaluated its performances. We showed promising results: after training with only virtual data, the network achieves 51.8 mAP on real data and 87.2 mAP on virtual data. After applying Domain Adaptation, the network achieves 76.2 mAP on real data and 73.3 mAP on virtual data.
File