This repository contains the data and code for the following paper:
**Can NLI Models Verify QA Systems' Predictions? **
Jifan Chen, Eunsol Choi, Greg Durrett
EMNLP 2021 Findings
@article{chen2021can,
title={Can NLI Models Verify QA Systems' Predictions?},
author={Chen, Jifan and Choi, Eunsol and Durrett, Greg},
journal={EMNLP Findings},
year={2021}
}
The NLI data converted from QA datasets through our pipeline described in the paper can be found here
The data files are formatted as jsonlines; each example is described as the following:
Field | Description |
---|---|
example_id |
Example ID |
title_text |
Title of the Wikipedia page of the example, could be NONE |
paragraph_text |
Paragraph containing the answer |
question_text |
Question |
answer_text |
Answer of the question |
answer_sent_text |
Sentence containing the answer |
decontext_answer_sent_text |
Decontextualized sentence containing the answer |
question_statement_text |
Declarative version of the question by combining the answer |
answer_scores |
Top 5 Answer score computed by the QA(BERT-joint) model |
is_correct |
Whether the answer is correct |
answer_sent_text |
Sentence containing the answer |
git clone https://github.com/jifan-chen/QA-Verification-Via-NLI.git
Install the dependencies by running
pip install -r requirements.txt
See README in seq2seq_converter.
To run the pre-trained NQ-NLI model, simply run bash scripts/run_nq_nli_prediction.sh
. The pre-trained Roberta-based model and dataset will be automatically downloaded and the predictions will be saved as ${dataset_path}-predictions.csv and ${dataset_path}-predictions.json.
Please contact at jfchen@cs.utexas.edu
if you have any questions.