Home > Research > Publications & Outputs > Explainable-by-Design Deep Learning

Electronic data

Text available via DOI:

View graph of relations

Explainable-by-Design Deep Learning

Research output: ThesisDoctoral Thesis

Published
Publication date15/08/2022
Number of pages160
QualificationPhD
Awarding Institution
Supervisors/Advisors
Thesis sponsors
  • Ford Motor Co, Ford Motor Company, Ford Res & Innovat Ctr
Award date22/03/2022
Publisher
  • Lancaster University
<mark>Original language</mark>English

Abstract

Machine learning, and more specifically, deep learning, have attracted the attention of media and the broader public in the last decade due to its potential to revolutionize industries, public services, and society. Deep learning achieved or even surpassed human experts’ performance in terms of accuracy for different challenging problems such as image recognition, speech, and language translation. However, deep learning models are often characterized as a “black box” as these models are composed of many millions of parameters, which are extremely difficult to interpret by specialists. Complex “black box” models can easily fool users unable to inspect the algorithm’s decision, which can lead to dangerous or catastrophic events. Therefore, auditable explainable AI approaches are crucial for developing safe systems, complying with regulations, and accepting this new technology within society. This thesis tries to answer the following research question: Is it possible to provide an approach that has a performance compared to a Deep Learning and the same time has a transparent structure (non-black box)? To this end, it introduces a novel framework of explainable- by-design Deep Learning architectures that offers transparency and high accuracy, helping humans understand why a particular machine decision has been reached and whether or not it is trustworthy. Moreover, the proposed prototype-based framework has a flexible structure that allows the unsupervised detection of new classes and situations. The approaches proposed in thesis have been applied to multiple use cases, including image classification, fairness, deep recursive learning interpretation, and novelty detection.