Interpretability and transparency in artificial intelligence

Artificial Intelligence (AI) systems are frequently thought of as opaque, meaning their performance or logic is thought to be inaccessible or incomprehensible to human observers. Models can consist of millions of features connected in a complex web of dependent behaviours. Conveying this internal st...

Full description

Bibliographic Details
Main Author: Mittelstadt, B
Other Authors: Véliz, C
Format: Book section
Language:English
Published: Oxford University Press 2022
_version_ 1797111757697712128
author Mittelstadt, B
author2 Véliz, C
author_facet Véliz, C
Mittelstadt, B
author_sort Mittelstadt, B
collection OXFORD
description Artificial Intelligence (AI) systems are frequently thought of as opaque, meaning their performance or logic is thought to be inaccessible or incomprehensible to human observers. Models can consist of millions of features connected in a complex web of dependent behaviours. Conveying this internal state and dependencies in a humanly comprehensible way is extremely challenging. Explaining the functionality and behaviour of AI systems in a meaningful and useful way to people designing, operating, regulating, or affected by their outputs is a complex technical, philosophical, and ethical project. Despite this complexity, principles citing ‘transparency’ or ‘interpretability’ are commonly found in ethical and regulatory frameworks addressing technology. This chapter provides an overview of these concepts and methods design to explain how AI works. After reviewing key concepts and terminology, two sets of methods are examined: (1) interpretability methods designed to explain and approximate AI functionality and behaviour; and (2) transparency frameworks meant to help assess and provide information about the development, governance, and potential impact of training datasets, models, and specific applications. These methods are analysed in the context of prior work on explanations in the philosophy of science. The chapter closes by introducing a framework of criteria to evaluate the quality and utility of methods in explainable AI (XAI) and to clarify the open challenges facing the field.
first_indexed 2024-03-07T08:13:21Z
format Book section
id oxford-uuid:e8c4eb25-b433-446d-b3fc-f7dfbb6e8fb3
institution University of Oxford
language English
last_indexed 2024-03-07T08:13:21Z
publishDate 2022
publisher Oxford University Press
record_format dspace
spelling oxford-uuid:e8c4eb25-b433-446d-b3fc-f7dfbb6e8fb32023-12-07T10:29:01ZInterpretability and transparency in artificial intelligenceBook sectionhttp://purl.org/coar/resource_type/c_1843uuid:e8c4eb25-b433-446d-b3fc-f7dfbb6e8fb3EnglishSymplectic ElementsOxford University Press2022Mittelstadt, BVéliz, CArtificial Intelligence (AI) systems are frequently thought of as opaque, meaning their performance or logic is thought to be inaccessible or incomprehensible to human observers. Models can consist of millions of features connected in a complex web of dependent behaviours. Conveying this internal state and dependencies in a humanly comprehensible way is extremely challenging. Explaining the functionality and behaviour of AI systems in a meaningful and useful way to people designing, operating, regulating, or affected by their outputs is a complex technical, philosophical, and ethical project. Despite this complexity, principles citing ‘transparency’ or ‘interpretability’ are commonly found in ethical and regulatory frameworks addressing technology. This chapter provides an overview of these concepts and methods design to explain how AI works. After reviewing key concepts and terminology, two sets of methods are examined: (1) interpretability methods designed to explain and approximate AI functionality and behaviour; and (2) transparency frameworks meant to help assess and provide information about the development, governance, and potential impact of training datasets, models, and specific applications. These methods are analysed in the context of prior work on explanations in the philosophy of science. The chapter closes by introducing a framework of criteria to evaluate the quality and utility of methods in explainable AI (XAI) and to clarify the open challenges facing the field.
spellingShingle Mittelstadt, B
Interpretability and transparency in artificial intelligence
title Interpretability and transparency in artificial intelligence
title_full Interpretability and transparency in artificial intelligence
title_fullStr Interpretability and transparency in artificial intelligence
title_full_unstemmed Interpretability and transparency in artificial intelligence
title_short Interpretability and transparency in artificial intelligence
title_sort interpretability and transparency in artificial intelligence
work_keys_str_mv AT mittelstadtb interpretabilityandtransparencyinartificialintelligence