Research

Narrative Question Answering with Cutting-Edge Open-Domain QA Techniques: A Comprehensive Study

TACL

Authors

  • Xiangyang Mou
  • Chenghao Yang, Mo Yu
  • Bingsheng Yao
  • Xiaoxiao Guo
  • Saloni Potdar
  • Hui Su

Published on

06/07/2021

Recent advancements in open-domain question answering (ODQA), i.e., finding answers from large open-domain corpus like Wikipedia, have led to human-level performance on many datasets. However, progress in QA over book stories (Book QA) lags behind despite its similar task formulation to ODQA. This work provides a comprehensive and quantitative analysis about the difficulty of Book QA: (1) We benchmark the research on the NarrativeQA dataset with extensive experiments with cutting-edge ODQA techniques. This quantifies the challenges Book QA poses, as well as advances the published state-of-the-art with a ∼7\% absolute improvement on Rouge-L. (2) We further analyze the detailed challenges in Book QA through human studies. Our findings indicate that the event-centric questions dominate this task, which exemplifies the inability of existing QA models to handle event-oriented scenarios.

This paper has been published at TACL 2021

Please cite our work using the BibTeX below.

@misc{mou2021narrative,
      title={Narrative Question Answering with Cutting-Edge Open-Domain QA Techniques: A Comprehensive Study}, 
      author={Xiangyang Mou and Chenghao Yang and Mo Yu and Bingsheng Yao and Xiaoxiao Guo and Saloni Potdar and Hui Su},
      year={2021},
      eprint={2106.03826},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
Close Modal