Home repository for the dataset introduced in ACORN: Aspect-wise Commonsense Reasoning Explanation Evaluation. ACORN contains 3,500 human-written and LLM-generated explanations with aspect-wise quality ratings given by humans.
Also available on 🤗HuggingFace here!
The entire dataset is contained in ACORN.jsonl
. Each row consists of an explanation, related information, aggregated (majority-voted) ratings, and the full set of individual worker ratings.
Basic fields:
question
question textchoices
list of answer choiceslabel
correct answer indexexplanation
explanation textvoted_ratings
majority-voted ratingsworker_ratings
all worker ratings, saved as a dictionary of dictionaries (worker id → rating dict).
→ See Additional fields for the full list of fields.
Explanation quality is subjective and can depend on the intended use. Our choice includes both a general rating and fine-grained aspects of explanation quality assuming an ideal of fluent, sufficient, minimal, and contrastive explanations.
ACORN contains a blend of explanations from several sources. See Section 2.2 in the paper for a more detailed overview.
In addition to the fields listed in Files, the dataset contains the following information.
id
test sample IDq_id
original question IDe_id
original explanation IDq_source
question source (Commonsense QA or Balanced COPA)e_source
explanation source (→ Sources)triples
triple-form explanation (COPA-SSE only)postivies
,negatives
positive and negative statements (ECQA only)
If you use this dataset, please consider citing the following work.
@article{brassard2024acorn,
title = {ACORN: Aspect-wise Commonsense Reasoning Explanation Evaluation},
author = {Ana Brassard and Benjamin Heinzerling and Keito Kudo and Keisuke Sakaguchi and Kentaro Inui},
year = {2024},
journal = {arXiv preprint arXiv: 2405.04818}
}