A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences
Main Author: | |
---|---|
Publication Date: | 2023 |
Other Authors: | , , , , , , , , , , , , , , |
Format: | Article |
Language: | eng |
Source: | Repositórios Científicos de Acesso Aberto de Portugal (RCAAP) |
Download full: | https://hdl.handle.net/10316/113753 https://doi.org/10.1007/s10462-022-10256-8 |
Summary: | Since its emergence in the 1960s, Artificial Intelligence (AI) has grown to conquer many technology products and their fields of application. Machine learning, as a major part of the current AI solutions, can learn from the data and through experience to reach high performance on various tasks. This growing success of AI algorithms has led to a need for interpretability to understand opaque models such as deep neural networks. Various requirements have been raised from different domains, together with numerous tools to debug, justify outcomes, and establish the safety, fairness and reliability of the models. This variety of tasks has led to inconsistencies in the terminology with, for instance, terms such as interpretable, explainable and transparent being often used interchangeably in methodology papers. These words, however, convey different meanings and are "weighted" differently across domains, for example in the technical and social sciences. In this paper, we propose an overarching terminology of interpretability of AI systems that can be referred to by the technical developers as much as by the social sciences community to pursue clarity and efficiency in the definition of regulations for ethical and reliable AI development. We show how our taxonomy and definition of interpretable AI differ from the ones in previous research and how they apply with high versatility to several domains and use cases, proposing a-highly needed-standard for the communication among interdisciplinary areas of AI. |
id |
RCAP_84d6c93dcac93b4fea3558c50922c364 |
---|---|
oai_identifier_str |
oai:estudogeral.uc.pt:10316/113753 |
network_acronym_str |
RCAP |
network_name_str |
Repositórios Científicos de Acesso Aberto de Portugal (RCAAP) |
repository_id_str |
https://opendoar.ac.uk/repository/7160 |
spelling |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciencesInterpretabilityExplainable artificial intelligenceMachine learningSince its emergence in the 1960s, Artificial Intelligence (AI) has grown to conquer many technology products and their fields of application. Machine learning, as a major part of the current AI solutions, can learn from the data and through experience to reach high performance on various tasks. This growing success of AI algorithms has led to a need for interpretability to understand opaque models such as deep neural networks. Various requirements have been raised from different domains, together with numerous tools to debug, justify outcomes, and establish the safety, fairness and reliability of the models. This variety of tasks has led to inconsistencies in the terminology with, for instance, terms such as interpretable, explainable and transparent being often used interchangeably in methodology papers. These words, however, convey different meanings and are "weighted" differently across domains, for example in the technical and social sciences. In this paper, we propose an overarching terminology of interpretability of AI systems that can be referred to by the technical developers as much as by the social sciences community to pursue clarity and efficiency in the definition of regulations for ethical and reliable AI development. We show how our taxonomy and definition of interpretable AI differ from the ones in previous research and how they apply with high versatility to several domains and use cases, proposing a-highly needed-standard for the communication among interdisciplinary areas of AI.Springer Nature2023info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articlehttps://hdl.handle.net/10316/113753https://hdl.handle.net/10316/113753https://doi.org/10.1007/s10462-022-10256-8eng0269-2821Graziani, MaraDutkiewicz, LidiaCalvaresi, DavideAmorim, José PereiraYordanova, KaterinaVered, MorNair, RahulAbreu, Pedro HenriquesBlanke, TobiasPulignano, ValeriaPrior, John O.Lauwaert, LodeReijers, WesselDepeursinge, AdrienAndrearczyk, VincentMüller, Henninginfo:eu-repo/semantics/openAccessreponame:Repositórios Científicos de Acesso Aberto de Portugal (RCAAP)instname:FCCN, serviços digitais da FCT – Fundação para a Ciência e a Tecnologiainstacron:RCAAP2024-02-29T12:00:17Zoai:estudogeral.uc.pt:10316/113753Portal AgregadorONGhttps://www.rcaap.pt/oai/openaireinfo@rcaap.ptopendoar:https://opendoar.ac.uk/repository/71602025-05-29T06:06:37.030998Repositórios Científicos de Acesso Aberto de Portugal (RCAAP) - FCCN, serviços digitais da FCT – Fundação para a Ciência e a Tecnologiafalse |
dc.title.none.fl_str_mv |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences |
title |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences |
spellingShingle |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences Graziani, Mara Interpretability Explainable artificial intelligence Machine learning |
title_short |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences |
title_full |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences |
title_fullStr |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences |
title_full_unstemmed |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences |
title_sort |
A global taxonomy of interpretable AI: unifying the terminology for the technical and social sciences |
author |
Graziani, Mara |
author_facet |
Graziani, Mara Dutkiewicz, Lidia Calvaresi, Davide Amorim, José Pereira Yordanova, Katerina Vered, Mor Nair, Rahul Abreu, Pedro Henriques Blanke, Tobias Pulignano, Valeria Prior, John O. Lauwaert, Lode Reijers, Wessel Depeursinge, Adrien Andrearczyk, Vincent Müller, Henning |
author_role |
author |
author2 |
Dutkiewicz, Lidia Calvaresi, Davide Amorim, José Pereira Yordanova, Katerina Vered, Mor Nair, Rahul Abreu, Pedro Henriques Blanke, Tobias Pulignano, Valeria Prior, John O. Lauwaert, Lode Reijers, Wessel Depeursinge, Adrien Andrearczyk, Vincent Müller, Henning |
author2_role |
author author author author author author author author author author author author author author author |
dc.contributor.author.fl_str_mv |
Graziani, Mara Dutkiewicz, Lidia Calvaresi, Davide Amorim, José Pereira Yordanova, Katerina Vered, Mor Nair, Rahul Abreu, Pedro Henriques Blanke, Tobias Pulignano, Valeria Prior, John O. Lauwaert, Lode Reijers, Wessel Depeursinge, Adrien Andrearczyk, Vincent Müller, Henning |
dc.subject.por.fl_str_mv |
Interpretability Explainable artificial intelligence Machine learning |
topic |
Interpretability Explainable artificial intelligence Machine learning |
description |
Since its emergence in the 1960s, Artificial Intelligence (AI) has grown to conquer many technology products and their fields of application. Machine learning, as a major part of the current AI solutions, can learn from the data and through experience to reach high performance on various tasks. This growing success of AI algorithms has led to a need for interpretability to understand opaque models such as deep neural networks. Various requirements have been raised from different domains, together with numerous tools to debug, justify outcomes, and establish the safety, fairness and reliability of the models. This variety of tasks has led to inconsistencies in the terminology with, for instance, terms such as interpretable, explainable and transparent being often used interchangeably in methodology papers. These words, however, convey different meanings and are "weighted" differently across domains, for example in the technical and social sciences. In this paper, we propose an overarching terminology of interpretability of AI systems that can be referred to by the technical developers as much as by the social sciences community to pursue clarity and efficiency in the definition of regulations for ethical and reliable AI development. We show how our taxonomy and definition of interpretable AI differ from the ones in previous research and how they apply with high versatility to several domains and use cases, proposing a-highly needed-standard for the communication among interdisciplinary areas of AI. |
publishDate |
2023 |
dc.date.none.fl_str_mv |
2023 |
dc.type.status.fl_str_mv |
info:eu-repo/semantics/publishedVersion |
dc.type.driver.fl_str_mv |
info:eu-repo/semantics/article |
format |
article |
status_str |
publishedVersion |
dc.identifier.uri.fl_str_mv |
https://hdl.handle.net/10316/113753 https://hdl.handle.net/10316/113753 https://doi.org/10.1007/s10462-022-10256-8 |
url |
https://hdl.handle.net/10316/113753 https://doi.org/10.1007/s10462-022-10256-8 |
dc.language.iso.fl_str_mv |
eng |
language |
eng |
dc.relation.none.fl_str_mv |
0269-2821 |
dc.rights.driver.fl_str_mv |
info:eu-repo/semantics/openAccess |
eu_rights_str_mv |
openAccess |
dc.publisher.none.fl_str_mv |
Springer Nature |
publisher.none.fl_str_mv |
Springer Nature |
dc.source.none.fl_str_mv |
reponame:Repositórios Científicos de Acesso Aberto de Portugal (RCAAP) instname:FCCN, serviços digitais da FCT – Fundação para a Ciência e a Tecnologia instacron:RCAAP |
instname_str |
FCCN, serviços digitais da FCT – Fundação para a Ciência e a Tecnologia |
instacron_str |
RCAAP |
institution |
RCAAP |
reponame_str |
Repositórios Científicos de Acesso Aberto de Portugal (RCAAP) |
collection |
Repositórios Científicos de Acesso Aberto de Portugal (RCAAP) |
repository.name.fl_str_mv |
Repositórios Científicos de Acesso Aberto de Portugal (RCAAP) - FCCN, serviços digitais da FCT – Fundação para a Ciência e a Tecnologia |
repository.mail.fl_str_mv |
info@rcaap.pt |
_version_ |
1833602578918670336 |