ArXiv | 2021

Robustifying Multi-hop QA through Pseudo-Evidentiality Training

 
 
 
 

Abstract


This paper studies the bias problem of multihop question answering models, of answering correctly without correct reasoning. One way to robustify these models is by supervising to not only answer right, but also with right reasoning chains. An existing direction is to annotate reasoning chains to train models, requiring expensive additional annotations. In contrast, we propose a new approach to learn evidentiality, deciding whether the answer prediction is supported by correct evidences, without such annotations. Instead, we compare counterfactual changes in answer confidence with and without evidence sentences, to generate “pseudo-evidentiality” annotations. We validate our proposed model on an original set and challenge set in HotpotQA, showing that our method is accurate and robust in multi-hop reasoning.

Volume abs/2107.03242
Pages None
DOI 10.18653/v1/2021.acl-long.476
Language English
Journal ArXiv

Full Text