Explaining Deep Learning-Based Driver Models

Different systems based on Artificial Intelligence (AI) techniques are currently used in relevant areas such as healthcare, cybersecurity, natural language processing, and self-driving cars. However, many of these systems are developed with “black box” AI, which makes it difficult to explain how the...

Full description

Bibliographic Details
Main Authors: Maria Paz Sesmero Lorente, Elena Magán Lopez, Laura Alvarez Florez, Agapito Ledezma Espino, José Antonio Iglesias Martínez, Araceli Sanchis de Miguel
Format: Article
Language:English
Published: MDPI AG 2021-04-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/11/8/3321
_version_ 1797538520187797504
author Maria Paz Sesmero Lorente
Elena Magán Lopez
Laura Alvarez Florez
Agapito Ledezma Espino
José Antonio Iglesias Martínez
Araceli Sanchis de Miguel
author_facet Maria Paz Sesmero Lorente
Elena Magán Lopez
Laura Alvarez Florez
Agapito Ledezma Espino
José Antonio Iglesias Martínez
Araceli Sanchis de Miguel
author_sort Maria Paz Sesmero Lorente
collection DOAJ
description Different systems based on Artificial Intelligence (AI) techniques are currently used in relevant areas such as healthcare, cybersecurity, natural language processing, and self-driving cars. However, many of these systems are developed with “black box” AI, which makes it difficult to explain how they work. For this reason, explainability and interpretability are key factors that need to be taken into consideration in the development of AI systems in critical areas. In addition, different contexts produce different explainability needs which must be met. Against this background, Explainable Artificial Intelligence (XAI) appears to be able to address and solve this situation. In the field of automated driving, XAI is particularly needed because the level of automation is constantly increasing according to the development of AI techniques. For this reason, the field of XAI in the context of automated driving is of particular interest. In this paper, we propose the use of an explainable intelligence technique in the understanding of some of the tasks involved in the development of advanced driver-assistance systems (ADAS). Since ADAS assist drivers in driving functions, it is essential to know the reason for the decisions taken. In addition, trusted AI is the cornerstone of the confidence needed in this research area. Thus, due to the complexity and the different variables that are part of the decision-making process, this paper focuses on two specific tasks in this area: the detection of emotions and the distractions of drivers. The results obtained are promising and show the capacity of the explainable artificial techniques in the different tasks of the proposed environments.
first_indexed 2024-03-10T12:31:36Z
format Article
id doaj.art-1d8fc40d1510453db560af47b5a93b33
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-10T12:31:36Z
publishDate 2021-04-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-1d8fc40d1510453db560af47b5a93b332023-11-21T14:33:12ZengMDPI AGApplied Sciences2076-34172021-04-01118332110.3390/app11083321Explaining Deep Learning-Based Driver ModelsMaria Paz Sesmero Lorente0Elena Magán Lopez1Laura Alvarez Florez2Agapito Ledezma Espino3José Antonio Iglesias Martínez4Araceli Sanchis de Miguel5Computer Science Department, Universidad Carlos III de Madrid, 28911 Madrid, SpainComputer Science Department, Universidad Carlos III de Madrid, 28911 Madrid, SpainComputer Science Department, Universidad Carlos III de Madrid, 28911 Madrid, SpainComputer Science Department, Universidad Carlos III de Madrid, 28911 Madrid, SpainComputer Science Department, Universidad Carlos III de Madrid, 28911 Madrid, SpainComputer Science Department, Universidad Carlos III de Madrid, 28911 Madrid, SpainDifferent systems based on Artificial Intelligence (AI) techniques are currently used in relevant areas such as healthcare, cybersecurity, natural language processing, and self-driving cars. However, many of these systems are developed with “black box” AI, which makes it difficult to explain how they work. For this reason, explainability and interpretability are key factors that need to be taken into consideration in the development of AI systems in critical areas. In addition, different contexts produce different explainability needs which must be met. Against this background, Explainable Artificial Intelligence (XAI) appears to be able to address and solve this situation. In the field of automated driving, XAI is particularly needed because the level of automation is constantly increasing according to the development of AI techniques. For this reason, the field of XAI in the context of automated driving is of particular interest. In this paper, we propose the use of an explainable intelligence technique in the understanding of some of the tasks involved in the development of advanced driver-assistance systems (ADAS). Since ADAS assist drivers in driving functions, it is essential to know the reason for the decisions taken. In addition, trusted AI is the cornerstone of the confidence needed in this research area. Thus, due to the complexity and the different variables that are part of the decision-making process, this paper focuses on two specific tasks in this area: the detection of emotions and the distractions of drivers. The results obtained are promising and show the capacity of the explainable artificial techniques in the different tasks of the proposed environments.https://www.mdpi.com/2076-3417/11/8/3321Explainable Artificial Intelligence (XAI)advanced driver-assistance system (ADAS)automotive environmentbehavior driver recognitionemotions driver recognitionXRAI (Region-based saliency method)
spellingShingle Maria Paz Sesmero Lorente
Elena Magán Lopez
Laura Alvarez Florez
Agapito Ledezma Espino
José Antonio Iglesias Martínez
Araceli Sanchis de Miguel
Explaining Deep Learning-Based Driver Models
Applied Sciences
Explainable Artificial Intelligence (XAI)
advanced driver-assistance system (ADAS)
automotive environment
behavior driver recognition
emotions driver recognition
XRAI (Region-based saliency method)
title Explaining Deep Learning-Based Driver Models
title_full Explaining Deep Learning-Based Driver Models
title_fullStr Explaining Deep Learning-Based Driver Models
title_full_unstemmed Explaining Deep Learning-Based Driver Models
title_short Explaining Deep Learning-Based Driver Models
title_sort explaining deep learning based driver models
topic Explainable Artificial Intelligence (XAI)
advanced driver-assistance system (ADAS)
automotive environment
behavior driver recognition
emotions driver recognition
XRAI (Region-based saliency method)
url https://www.mdpi.com/2076-3417/11/8/3321
work_keys_str_mv AT mariapazsesmerolorente explainingdeeplearningbaseddrivermodels
AT elenamaganlopez explainingdeeplearningbaseddrivermodels
AT lauraalvarezflorez explainingdeeplearningbaseddrivermodels
AT agapitoledezmaespino explainingdeeplearningbaseddrivermodels
AT joseantonioiglesiasmartinez explainingdeeplearningbaseddrivermodels
AT aracelisanchisdemiguel explainingdeeplearningbaseddrivermodels