IDC: quantitative evaluation benchmark of interpretation methods for deep text classification models

Abstract Recent advances in deep neural networks have achieved outstanding success in natural language processing tasks. Interpretation methods that provide insight into the decision-making process of these models have received an influx of research attention because of the success and the black-box...

Full description

Bibliographic Details
Main Authors: Mohammed Khaleel, Lei Qi, Wallapak Tavanapong, Johnny Wong, Adisak Sukul, David A. M. Peterson
Format: Article
Language:English
Published: SpringerOpen 2022-03-01
Series:Journal of Big Data
Subjects:
Online Access:https://doi.org/10.1186/s40537-022-00583-6
_version_ 1819052470709518336
author Mohammed Khaleel
Lei Qi
Wallapak Tavanapong
Johnny Wong
Adisak Sukul
David A. M. Peterson
author_facet Mohammed Khaleel
Lei Qi
Wallapak Tavanapong
Johnny Wong
Adisak Sukul
David A. M. Peterson
author_sort Mohammed Khaleel
collection DOAJ
description Abstract Recent advances in deep neural networks have achieved outstanding success in natural language processing tasks. Interpretation methods that provide insight into the decision-making process of these models have received an influx of research attention because of the success and the black-box nature of the deep text classification models. Evaluation of these methods has been based on changes in classification accuracy or prediction confidence when removing important words identified by these methods. There are no measurements of the actual difference between the predicted important words and humans’ interpretation of ground truth because of the lack of interpretation ground truth. A large publicly available interpretation ground truth has the potential to advance the development of interpretation methods. Manual labeling important words for each document to create a large interpretation ground truth is very time-consuming. This paper presents (1) IDC, a new benchmark for quantitative evaluation of interpretation methods for deep text classification models, and (2) evaluation of six interpretation methods using the benchmark. The IDC benchmark consists of: (1) Three methods that generate three pseudo-interpretation ground truth datasets. (2) Three performance metrics: interpretation recall, interpretation precision, and Cohen’s kappa inter-agreement. Findings: IDC-generated interpretation ground truth agrees with human annotators on sampled movie reviews. IDC identifies Layer-wise Relevance Propagation and the gradient-by-input methods as the winning interpretation methods in this study.
first_indexed 2024-12-21T12:20:21Z
format Article
id doaj.art-5efa4849d9484c91a08f40a27c39b748
institution Directory Open Access Journal
issn 2196-1115
language English
last_indexed 2024-12-21T12:20:21Z
publishDate 2022-03-01
publisher SpringerOpen
record_format Article
series Journal of Big Data
spelling doaj.art-5efa4849d9484c91a08f40a27c39b7482022-12-21T19:04:20ZengSpringerOpenJournal of Big Data2196-11152022-03-019111410.1186/s40537-022-00583-6IDC: quantitative evaluation benchmark of interpretation methods for deep text classification modelsMohammed Khaleel0Lei Qi1Wallapak Tavanapong2Johnny Wong3Adisak Sukul4David A. M. Peterson5Department of Computer Science, Iowa State UniversityDepartment of Computer Science, Iowa State UniversityDepartment of Computer Science, Iowa State UniversityDepartment of Computer Science, Iowa State UniversityDepartment of Computer Science, Iowa State UniversityDepartment of Political Science, Iowa State UniversityAbstract Recent advances in deep neural networks have achieved outstanding success in natural language processing tasks. Interpretation methods that provide insight into the decision-making process of these models have received an influx of research attention because of the success and the black-box nature of the deep text classification models. Evaluation of these methods has been based on changes in classification accuracy or prediction confidence when removing important words identified by these methods. There are no measurements of the actual difference between the predicted important words and humans’ interpretation of ground truth because of the lack of interpretation ground truth. A large publicly available interpretation ground truth has the potential to advance the development of interpretation methods. Manual labeling important words for each document to create a large interpretation ground truth is very time-consuming. This paper presents (1) IDC, a new benchmark for quantitative evaluation of interpretation methods for deep text classification models, and (2) evaluation of six interpretation methods using the benchmark. The IDC benchmark consists of: (1) Three methods that generate three pseudo-interpretation ground truth datasets. (2) Three performance metrics: interpretation recall, interpretation precision, and Cohen’s kappa inter-agreement. Findings: IDC-generated interpretation ground truth agrees with human annotators on sampled movie reviews. IDC identifies Layer-wise Relevance Propagation and the gradient-by-input methods as the winning interpretation methods in this study.https://doi.org/10.1186/s40537-022-00583-6Machine learning interpretationNatural language processingPseudo interpretation ground truth
spellingShingle Mohammed Khaleel
Lei Qi
Wallapak Tavanapong
Johnny Wong
Adisak Sukul
David A. M. Peterson
IDC: quantitative evaluation benchmark of interpretation methods for deep text classification models
Journal of Big Data
Machine learning interpretation
Natural language processing
Pseudo interpretation ground truth
title IDC: quantitative evaluation benchmark of interpretation methods for deep text classification models
title_full IDC: quantitative evaluation benchmark of interpretation methods for deep text classification models
title_fullStr IDC: quantitative evaluation benchmark of interpretation methods for deep text classification models
title_full_unstemmed IDC: quantitative evaluation benchmark of interpretation methods for deep text classification models
title_short IDC: quantitative evaluation benchmark of interpretation methods for deep text classification models
title_sort idc quantitative evaluation benchmark of interpretation methods for deep text classification models
topic Machine learning interpretation
Natural language processing
Pseudo interpretation ground truth
url https://doi.org/10.1186/s40537-022-00583-6
work_keys_str_mv AT mohammedkhaleel idcquantitativeevaluationbenchmarkofinterpretationmethodsfordeeptextclassificationmodels
AT leiqi idcquantitativeevaluationbenchmarkofinterpretationmethodsfordeeptextclassificationmodels
AT wallapaktavanapong idcquantitativeevaluationbenchmarkofinterpretationmethodsfordeeptextclassificationmodels
AT johnnywong idcquantitativeevaluationbenchmarkofinterpretationmethodsfordeeptextclassificationmodels
AT adisaksukul idcquantitativeevaluationbenchmarkofinterpretationmethodsfordeeptextclassificationmodels
AT davidampeterson idcquantitativeevaluationbenchmarkofinterpretationmethodsfordeeptextclassificationmodels