Tesi etd-01262025-170550 |
Link copiato negli appunti
Tipo di tesi
Tesi di dottorato di ricerca
Autore
MARCONATO, EMANUELE
URN
etd-01262025-170550
Titolo
Learning Concepts with the Right Semantics: Reasoning Shortcuts and Human-Machine Alignment
Settore scientifico disciplinare
INF/01 - INFORMATICA
Corso di studi
DOTTORATO NAZIONALE IN INTELLIGENZA ARTIFICIALE
Relatori
tutor Prof. Passerini, Andrea
correlatore Prof. Teso, Stefano
correlatore Prof. Barra, Adriano
correlatore Prof. Teso, Stefano
correlatore Prof. Barra, Adriano
Parole chiave
- Alignment
- Causal Representation Learning
- Explainable AI
- Interpretability
- Neuro-Symbolic AI
- Shortcuts
Data inizio appello
18/02/2025
Consultabilità
Completa
Riassunto
Understanding the functioning of current AI models is an urgent open problem, due to the massive scale use of deep neural networks and their black-box nature. Several works address how to explain the behavior of AI models and greater interest is growing to provide explanations with high-level variables, often called concepts or symbols. Leveraging concepts as a vehicle for explaining AI models allows to discard irrelevant information and focus only on the semantic content of the data. This has the potential to make models more interpretable, and achieve higher trust in their decision-making process. One key open problem is how to learn concepts from data such that they possess the correct semantics. This thesis analyzes this prob- lem in depth, presenting two major contributions. The first is explaining and addressing pitfalls in learning the right concepts in the context of tasks that involve reasoning on them. These pitfalls are due to Reasoning Shortcuts, whereby models can leverage poor-quality concepts to attain correct predictions. The second contribution is establishing a formal framework to test the quality of concepts learned by the model, and successively presenting a class of models that boost concept quality by leveraging advanced representation learning techniques. Overall, the presented works contribute to further understanding the issues complicating provably learning the concepts from data and to designing more trustworthy AI models for future high-stakes applications.
File
Nome file | Dimensione |
---|---|
phd_thes...nal_2.pdf | 22.78 Mb |
Contatta l’autore |