Exploring explainable AI: category theory insights into machine learning algorithms
Explainable artificial intelligence (XAI) is a growing field that aims to increase the transparency and interpretability of machine learning (ML) models. The aim of this work is to use the categorical properties of learning algorithms in conjunction with the categorical perspective of the informatio...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IOP Publishing
2023-01-01
|
Series: | Machine Learning: Science and Technology |
Subjects: | |
Online Access: | https://doi.org/10.1088/2632-2153/ad1534 |
_version_ | 1797373894135382016 |
---|---|
author | Ares Fabregat-Hernández Javier Palanca Vicent Botti |
author_facet | Ares Fabregat-Hernández Javier Palanca Vicent Botti |
author_sort | Ares Fabregat-Hernández |
collection | DOAJ |
description | Explainable artificial intelligence (XAI) is a growing field that aims to increase the transparency and interpretability of machine learning (ML) models. The aim of this work is to use the categorical properties of learning algorithms in conjunction with the categorical perspective of the information in the datasets to give a framework for explainability. In order to achieve this, we are going to define the enriched categories, with decorated morphisms, $\pmb{\mathcal{Learn}}$ , $\pmb{\mathcal{Para}}$ and $\pmb{\mathcal{MNet}}$ of learners, parameterized functions, and neural networks over metric spaces respectively. The main idea is to encode information from the dataset via categorical methods, see how it propagates, and lastly, interpret the results thanks again to categorical (metric) information. This means that we can attach numerical (computable) information via enrichment to the structural information of the category. With this, we can translate theoretical information into parameters that are easily understandable. We will make use of different categories of enrichment to keep track of different kinds of information. That is, to see how differences in attributes of the data are modified by the algorithm to result in differences in the output to achieve better separation. In that way, the categorical framework gives us an algorithm to interpret what the learning algorithm is doing. Furthermore, since it is designed with generality in mind, it should be applicable in various different contexts. There are three main properties of category theory that help with the interpretability of ML models: formality, the existence of universal properties, and compositionality. The last property offers a way to combine smaller, simpler models that are easily understood to build larger ones. This is achieved by formally representing the structure of ML algorithms and information contained in the model. Finally, universal properties are a cornerstone of category theory. They help us characterize an object, not by its attributes, but by how it interacts with other objects. Thus, we can formally characterize an algorithm by how it interacts with the data. The main advantage of the framework is that it can unify under the same language different techniques used in XAI. Thus, using the same language and concepts we can describe a myriad of techniques and properties of ML algorithms, streamlining their explanation and making them easier to generalize and extrapolate. |
first_indexed | 2024-03-08T18:57:02Z |
format | Article |
id | doaj.art-908cdbd408bb4988be583ab7648d2048 |
institution | Directory Open Access Journal |
issn | 2632-2153 |
language | English |
last_indexed | 2024-03-08T18:57:02Z |
publishDate | 2023-01-01 |
publisher | IOP Publishing |
record_format | Article |
series | Machine Learning: Science and Technology |
spelling | doaj.art-908cdbd408bb4988be583ab7648d20482023-12-28T10:52:36ZengIOP PublishingMachine Learning: Science and Technology2632-21532023-01-014404506110.1088/2632-2153/ad1534Exploring explainable AI: category theory insights into machine learning algorithmsAres Fabregat-Hernández0https://orcid.org/0009-0005-5104-514XJavier Palanca1Vicent Botti2https://orcid.org/0000-0002-6507-2756Valencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politécnica de Valéncia , Camí de Vera s/n, 46022 Valencia, SpainValencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politécnica de Valéncia , Camí de Vera s/n, 46022 Valencia, SpainValencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politécnica de Valéncia , Camí de Vera s/n, 46022 Valencia, Spain; valgrAI (Valencian Graduate School and Research Network of Artificial Intelligence) , Valencia, SpainExplainable artificial intelligence (XAI) is a growing field that aims to increase the transparency and interpretability of machine learning (ML) models. The aim of this work is to use the categorical properties of learning algorithms in conjunction with the categorical perspective of the information in the datasets to give a framework for explainability. In order to achieve this, we are going to define the enriched categories, with decorated morphisms, $\pmb{\mathcal{Learn}}$ , $\pmb{\mathcal{Para}}$ and $\pmb{\mathcal{MNet}}$ of learners, parameterized functions, and neural networks over metric spaces respectively. The main idea is to encode information from the dataset via categorical methods, see how it propagates, and lastly, interpret the results thanks again to categorical (metric) information. This means that we can attach numerical (computable) information via enrichment to the structural information of the category. With this, we can translate theoretical information into parameters that are easily understandable. We will make use of different categories of enrichment to keep track of different kinds of information. That is, to see how differences in attributes of the data are modified by the algorithm to result in differences in the output to achieve better separation. In that way, the categorical framework gives us an algorithm to interpret what the learning algorithm is doing. Furthermore, since it is designed with generality in mind, it should be applicable in various different contexts. There are three main properties of category theory that help with the interpretability of ML models: formality, the existence of universal properties, and compositionality. The last property offers a way to combine smaller, simpler models that are easily understood to build larger ones. This is achieved by formally representing the structure of ML algorithms and information contained in the model. Finally, universal properties are a cornerstone of category theory. They help us characterize an object, not by its attributes, but by how it interacts with other objects. Thus, we can formally characterize an algorithm by how it interacts with the data. The main advantage of the framework is that it can unify under the same language different techniques used in XAI. Thus, using the same language and concepts we can describe a myriad of techniques and properties of ML algorithms, streamlining their explanation and making them easier to generalize and extrapolate.https://doi.org/10.1088/2632-2153/ad1534explainabilitycategory theoryLipschitz functionsYoneda embeddingcompositionality |
spellingShingle | Ares Fabregat-Hernández Javier Palanca Vicent Botti Exploring explainable AI: category theory insights into machine learning algorithms Machine Learning: Science and Technology explainability category theory Lipschitz functions Yoneda embedding compositionality |
title | Exploring explainable AI: category theory insights into machine learning algorithms |
title_full | Exploring explainable AI: category theory insights into machine learning algorithms |
title_fullStr | Exploring explainable AI: category theory insights into machine learning algorithms |
title_full_unstemmed | Exploring explainable AI: category theory insights into machine learning algorithms |
title_short | Exploring explainable AI: category theory insights into machine learning algorithms |
title_sort | exploring explainable ai category theory insights into machine learning algorithms |
topic | explainability category theory Lipschitz functions Yoneda embedding compositionality |
url | https://doi.org/10.1088/2632-2153/ad1534 |
work_keys_str_mv | AT aresfabregathernandez exploringexplainableaicategorytheoryinsightsintomachinelearningalgorithms AT javierpalanca exploringexplainableaicategorytheoryinsightsintomachinelearningalgorithms AT vicentbotti exploringexplainableaicategorytheoryinsightsintomachinelearningalgorithms |