Limitations of Feature Attribution in Long Text Classification of Standards

Authors

  • Katharina Beckh Fraunhofer IAIS Lamarr Institute for Machine Learning and Artificial Intelligence
  • Joann Rachel Jacob Fraunhofer IAIS
  • Adrian Seeliger Deutsches Institut für Normung e. V. (DIN)
  • Stefan Rüping Fraunhofer IAIS
  • Najmeh Mousavi Nejad Fraunhofer IAIS

DOI:

https://doi.org/10.1609/aaaiss.v4i1.31765

Abstract

Managing complex AI systems requires insight into a model's decision-making processes. Understanding how these systems arrive at their conclusions is essential for ensuring reliability. In the field of explainable natural language processing, many approaches have been developed and evaluated. However, experimental analysis of explainability for text classification has been largely constrained to short text and binary classification. In this applied work, we study explainability for a real-world task where the goal is to assess the technological suitability of standards. This prototypical use case is characterized by large documents, technical language, and a multi-label setting, making it a complex modeling challenge. We provide an analysis of approx. 1000 documents with human-annotated evidence. We then present experimental results with two explanation methods evaluating plausibility and runtime of explanations. We find that the average runtime for explanation generation is at least 5 minutes and that the model explanations do not overlap with the ground truth. These findings reveal limitations of current explanation methods. In a detailed discussion, we identify possible reasons and how to address them on three different dimensions: task, model and explanation method. We conclude with risks and recommendations for the use of feature attribution methods in similar settings.

Downloads

Published

2024-11-08

How to Cite

Beckh, K., Jacob, J. R., Seeliger, A., Rüping, S., & Mousavi Nejad, N. (2024). Limitations of Feature Attribution in Long Text Classification of Standards. Proceedings of the AAAI Symposium Series, 4(1), 10-17. https://doi.org/10.1609/aaaiss.v4i1.31765

Issue

Section

AI Trustworthiness and Risk Assessment for Challenging Contexts (ATRACC)