Transactions of the Association for Computational Linguistics (May 2023)

Less is More: Mitigate Spurious Correlations for Open-Domain Dialogue Response Generation Models by Causal Discovery

  • Tao Feng,
  • Lizhen Qu,
  • Gholamreza Haffari

DOI
https://doi.org/10.1162/tacl_a_00561
Journal volume & issue
Vol. 11
pp. 511 – 530

Abstract

Read online

AbstractIn this paper, we conduct the first study on spurious correlations for open-domain response generation models based on a corpus CGDialog curated by ourselves. The current models indeed suffer from spurious correlations and have a tendency to generate irrelevant and generic responses. Inspired by causal discovery algorithms, we propose a novel model-agnostic method for training and inference using a conditional independence classifier. The classifier is trained by a constrained self-training method, coined ConSTrain, to overcome data sparsity. The experimental results based on both human and automatic evaluation show that our method significantly outperforms the competitive baselines in terms of relevance, informativeness, and fluency.