An Attention Matrix for Every Decision: Faithfulness-based Arbitration Among Multiple Attention-Based Interpretations of Transformers in Text Classification (original) (raw)
Related papers
Toward Practical Usage of the Attention Mechanism as a Tool for Interpretability
IEEE Access
More Identifiable yet Equally Performant Transformers for Text Classification
ArXiv, 2021
On the Lack of Robust Interpretability of Neural Text Classifiers
2021
Attention Interpretability Across NLP Tasks
ArXiv, 2019
End-to-End Transformer-Based Models in Textual-Based NLP
AI
Tracr: Compiled Transformers as a Laboratory for Interpretability
arXiv (Cornell University), 2023
Interpreting Convolutional Networks Trained on Textual Data
2021
arXiv (Cornell University), 2020
Looking Deeper into Deep Learning Model: Attribution-based Explanations of TextCNN
ArXiv, 2018
Perturbing Inputs for Fragile Interpretations in Deep Natural Language Processing
2021
Extractive Summarization for Explainable Sentiment Analysis using Transformers
2021
AbductionRules: Training Transformers to Explain Unexpected Inputs
Findings of the Association for Computational Linguistics: ACL 2022
TransICD: Transformer Based Code-Wise Attention Model for Explainable ICD Coding
Artificial Intelligence in Medicine, 2021
BaSFormer: A Balanced Sparsity Regularized Attention Network for Transformer
2023
Hierarchical Interpretation of Neural Text Classification
2022
Stress Test Evaluation of Transformer-based Models in Natural Language Understanding Tasks
arXiv (Cornell University), 2020
DoLFIn: Distributions over Latent Features for Interpretability
Proceedings of the 28th International Conference on Computational Linguistics, 2020
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021
IDC: Quantitative Evaluation Benchmark of Interpretation Methods for Deep Text Classification Models
2021
Commonsense Statements Identification and Explanation with Transformer-based Encoders
Proceedings of Deep Learning Inside Out (DeeLIO): The First Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, 2020
A Study of the Plausibility of Attention between RNN Encoders in Natural Language Inference
2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), 2021
Investigation of Transformer-based Latent Attention Models for Neural Machine Translation
2020
Revisiting Transformer-based Models for Long Document Classification
ArXiv, 2022
COMPARATIVE ANALYSIS OF TRANSFORMER BASED LANGUAGE MODELS
Computer Science & Information Technology (CS & IT) Computer Science Conference Proceedings (CSCP)
Enhancing Attention’s Explanation Using Interpretable Tsetlin Machine
Algorithms
Overview of the Transformer-based Models for NLP Tasks
Proceedings of the 2020 Federated Conference on Computer Science and Information Systems, 2020
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019
RCMHA: Relative Convolutional Multi-Head Attention for Natural Language Modelling
arXiv (Cornell University), 2023
Local Interpretations for Explainable Natural Language Processing: A Survey
arXiv (Cornell University), 2021
Exploring the Role of Transformers in NLP: From BERT to GPT-3
IRJET, 2023
Hierarchical Transformers for Long Document Classification
2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), 2019
Proceedings of the 36th Annual ACM Symposium on Applied Computing, 2021
An Introductory Survey on Attention Mechanisms in NLP Problems