Analyzing Semantic Faithfulness of Language Models via Input Intervention on Question Answering

Authors

  • Akshay Chaturvedi ANITI
  • Swarnadeep Bhar Institut de Recherche en Informatique de Toulouse
  • Soumadeep Saha the Indian Statistical Institute
  • Utpal Garain Indian Statistical Insititute
  • Nicholas Asher IRIT, Centre National de Recherche Scientifique, France

Abstract

Transformer-based language models have been shown to be highly effective for several NLP tasks. In this article, we consider three transformer models, BERT, RoBERTa, and XLNet, in both small and large versions, and investigate how faithful their representations are with respect to the semantic content of texts. We formalize a notion of semantic faithfulness, in which the semantic content of a text should causally figure in a model’s inferences in question answering. We then test this notion by observing a model’s behavior on answering questions about a story after performing two novel semantic interventions—deletion intervention and negation intervention. While transformer models achieve high performance on standard question answering tasks, we show that they fail to be semantically faithful once we perform these interventions for a significant number of cases (∼ 50% for deletion intervention, and ∼ 20% drop in accuracy for negation intervention). We then propose an intervention-based training regime that can mitigate the undesirable effects for deletion intervention by a significant margin (from ∼ 50% to ∼ 6%). We analyze the inner-workings of the models to better understand the effectiveness of intervention-based training for deletion intervention. But we show that this training does not attenuate other aspects of semantic unfaithfulness such as the models’ inability to deal with negation intervention or to capture the predicate–argument structure of texts. We also test InstructGPT, via prompting, for its ability to handle the two interventions and to capture predicate–argument structure. While InstructGPT models do achieve very high performance on predicate–argument structure task, they fail to respond adequately to our deletion and negation interventions.

Author Biographies

  • Akshay Chaturvedi, ANITI
    Akshay Chaturvedi is a post doctoral researcher in the interdisciplinary AI institute ANITI of Toulouse
  • Swarnadeep Bhar, Institut de Recherche en Informatique de Toulouse
    Swarnadeep Bhar is a PhD student at IRIT and participates in ANITI.
  • Soumadeep Saha, the Indian Statistical Institute
    Soumadeep Saha is a PhD student at the Indian Statistical Institute
  • Utpal Garain, Indian Statistical Insititute
    Utpal Garain is professor at the Indian Statistical Institute
  • Nicholas Asher, IRIT, Centre National de Recherche Scientifique, France
    Nicholas Asher is a senior director of research at the Centre Nationale de Recherche Scientifique in France

Published

2024-09-02