Abstract

We present the Stanford Question Answering Dataset (SQuAD), a new reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on a set of Wikipedia articles, where the answer to each question is a segment of text from the corresponding reading passage. We analyze the dataset to understand the types of reasoning required to answer the questions, leaning heavily on dependency and constituency trees. We build a strong logistic regression model, which achieves an F1 score of 51.0%, a significant improvement over a simple baseline (20%). However, human performance (86.8%) is much higher, indicating that the dataset presents a good challenge problem for future research. The dataset is freely available at https://stanford-qa.com

Keywords

Computer scienceComprehensionNatural language processingArtificial intelligenceInformation retrievalProgramming language

Related Publications

Publication Info

Year
2016
Type
article
Pages
2383-2392
Citations
6054
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

6054
OpenAlex
1705
Influential
2390
CrossRef

Cite This

Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev et al. (2016). SQuAD: 100,000+ Questions for Machine Comprehension of Text. Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing , 2383-2392. https://doi.org/10.18653/v1/d16-1264

Identifiers

DOI
10.18653/v1/d16-1264
arXiv
1606.05250

Data Quality

Data completeness: 84%