Exploring explainable AI: category theory insights into machine learning algorithms

Explainable artificial intelligence (XAI) is a growing field that aims to increase the transparency and interpretability of machine learning (ML) models. The aim of this work is to use the categorical properties of learning algorithms in conjunction with the categorical perspective of the informatio...

Full description

Bibliographic Details
Main Authors: Ares Fabregat-Hernández, Javier Palanca, Vicent Botti
Format: Article
Language:English
Published: IOP Publishing 2023-01-01
Series:Machine Learning: Science and Technology
Subjects:
Online Access:https://doi.org/10.1088/2632-2153/ad1534
_version_ 1797373894135382016
author Ares Fabregat-Hernández
Javier Palanca
Vicent Botti
author_facet Ares Fabregat-Hernández
Javier Palanca
Vicent Botti
author_sort Ares Fabregat-Hernández
collection DOAJ
description Explainable artificial intelligence (XAI) is a growing field that aims to increase the transparency and interpretability of machine learning (ML) models. The aim of this work is to use the categorical properties of learning algorithms in conjunction with the categorical perspective of the information in the datasets to give a framework for explainability. In order to achieve this, we are going to define the enriched categories, with decorated morphisms, $\pmb{\mathcal{Learn}}$ , $\pmb{\mathcal{Para}}$ and $\pmb{\mathcal{MNet}}$ of learners, parameterized functions, and neural networks over metric spaces respectively. The main idea is to encode information from the dataset via categorical methods, see how it propagates, and lastly, interpret the results thanks again to categorical (metric) information. This means that we can attach numerical (computable) information via enrichment to the structural information of the category. With this, we can translate theoretical information into parameters that are easily understandable. We will make use of different categories of enrichment to keep track of different kinds of information. That is, to see how differences in attributes of the data are modified by the algorithm to result in differences in the output to achieve better separation. In that way, the categorical framework gives us an algorithm to interpret what the learning algorithm is doing. Furthermore, since it is designed with generality in mind, it should be applicable in various different contexts. There are three main properties of category theory that help with the interpretability of ML models: formality, the existence of universal properties, and compositionality. The last property offers a way to combine smaller, simpler models that are easily understood to build larger ones. This is achieved by formally representing the structure of ML algorithms and information contained in the model. Finally, universal properties are a cornerstone of category theory. They help us characterize an object, not by its attributes, but by how it interacts with other objects. Thus, we can formally characterize an algorithm by how it interacts with the data. The main advantage of the framework is that it can unify under the same language different techniques used in XAI. Thus, using the same language and concepts we can describe a myriad of techniques and properties of ML algorithms, streamlining their explanation and making them easier to generalize and extrapolate.
first_indexed 2024-03-08T18:57:02Z
format Article
id doaj.art-908cdbd408bb4988be583ab7648d2048
institution Directory Open Access Journal
issn 2632-2153
language English
last_indexed 2024-03-08T18:57:02Z
publishDate 2023-01-01
publisher IOP Publishing
record_format Article
series Machine Learning: Science and Technology
spelling doaj.art-908cdbd408bb4988be583ab7648d20482023-12-28T10:52:36ZengIOP PublishingMachine Learning: Science and Technology2632-21532023-01-014404506110.1088/2632-2153/ad1534Exploring explainable AI: category theory insights into machine learning algorithmsAres Fabregat-Hernández0https://orcid.org/0009-0005-5104-514XJavier Palanca1Vicent Botti2https://orcid.org/0000-0002-6507-2756Valencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politécnica de Valéncia , Camí de Vera s/n, 46022 Valencia, SpainValencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politécnica de Valéncia , Camí de Vera s/n, 46022 Valencia, SpainValencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politécnica de Valéncia , Camí de Vera s/n, 46022 Valencia, Spain; valgrAI (Valencian Graduate School and Research Network of Artificial Intelligence) , Valencia, SpainExplainable artificial intelligence (XAI) is a growing field that aims to increase the transparency and interpretability of machine learning (ML) models. The aim of this work is to use the categorical properties of learning algorithms in conjunction with the categorical perspective of the information in the datasets to give a framework for explainability. In order to achieve this, we are going to define the enriched categories, with decorated morphisms, $\pmb{\mathcal{Learn}}$ , $\pmb{\mathcal{Para}}$ and $\pmb{\mathcal{MNet}}$ of learners, parameterized functions, and neural networks over metric spaces respectively. The main idea is to encode information from the dataset via categorical methods, see how it propagates, and lastly, interpret the results thanks again to categorical (metric) information. This means that we can attach numerical (computable) information via enrichment to the structural information of the category. With this, we can translate theoretical information into parameters that are easily understandable. We will make use of different categories of enrichment to keep track of different kinds of information. That is, to see how differences in attributes of the data are modified by the algorithm to result in differences in the output to achieve better separation. In that way, the categorical framework gives us an algorithm to interpret what the learning algorithm is doing. Furthermore, since it is designed with generality in mind, it should be applicable in various different contexts. There are three main properties of category theory that help with the interpretability of ML models: formality, the existence of universal properties, and compositionality. The last property offers a way to combine smaller, simpler models that are easily understood to build larger ones. This is achieved by formally representing the structure of ML algorithms and information contained in the model. Finally, universal properties are a cornerstone of category theory. They help us characterize an object, not by its attributes, but by how it interacts with other objects. Thus, we can formally characterize an algorithm by how it interacts with the data. The main advantage of the framework is that it can unify under the same language different techniques used in XAI. Thus, using the same language and concepts we can describe a myriad of techniques and properties of ML algorithms, streamlining their explanation and making them easier to generalize and extrapolate.https://doi.org/10.1088/2632-2153/ad1534explainabilitycategory theoryLipschitz functionsYoneda embeddingcompositionality
spellingShingle Ares Fabregat-Hernández
Javier Palanca
Vicent Botti
Exploring explainable AI: category theory insights into machine learning algorithms
Machine Learning: Science and Technology
explainability
category theory
Lipschitz functions
Yoneda embedding
compositionality
title Exploring explainable AI: category theory insights into machine learning algorithms
title_full Exploring explainable AI: category theory insights into machine learning algorithms
title_fullStr Exploring explainable AI: category theory insights into machine learning algorithms
title_full_unstemmed Exploring explainable AI: category theory insights into machine learning algorithms
title_short Exploring explainable AI: category theory insights into machine learning algorithms
title_sort exploring explainable ai category theory insights into machine learning algorithms
topic explainability
category theory
Lipschitz functions
Yoneda embedding
compositionality
url https://doi.org/10.1088/2632-2153/ad1534
work_keys_str_mv AT aresfabregathernandez exploringexplainableaicategorytheoryinsightsintomachinelearningalgorithms
AT javierpalanca exploringexplainableaicategorytheoryinsightsintomachinelearningalgorithms
AT vicentbotti exploringexplainableaicategorytheoryinsightsintomachinelearningalgorithms