Name | Description | Size | License |
---|---|---|---|
AI2 Science Exams | Elementary science questions from US state and regional science exams | 170 multi-state and 108 4th grade questions. | Creative Commons ShareAlike |
DeepMind Daily Mail | Collection of news articles and corresponding cloze queries |
196,961 documents
879,450 queries |
Apache |
DeepMind CNN | Collection of news articles and corresponding cloze queries |
90,266 documents
380,298 queries |
Apache |
Facebook bAbI SimpleQuestions | Generated data for 20 tasks testing specific aspects of text understanding and reasoning | 108,442 questions | BSD |
Facebook bAbI Chidlren's Book Test | Text passages and corresponding questions drawn from Project Gutenberg Children's books | 669,343 training questions 8,000 dev questions 10,000 test questions |
BSD |
MCTest | Machine comprehension of short stories | 660 stories, 4 questions per story |
Proprietary |
SQuAD: Stanford Question Answering Dataset | Crowdsourced question and answers with corresponding Wikipedia passages | 107,785 question-answer pairs on 536 articles | |
Todai Robot | Exams for entrance into the University of Tokyo | N | UNK |
-
MCTest is a data set created by Microsoft Research and consists of 660 short, fictional stories, each with an associated set of multiple-choice questions. Each story and corresponding question set was created via crowdsourcing. The participants in each story are unique, and each story independent of the next. The data is split into MC 160, which has been manually curated to correct errors occuring in the crowdsourced data, and the larger MC 500.
As shown below, each story is accompanied by four questions. According to the crowdsourcing guidelines, at least two of the four questions should require multi-sentence reasoning.
James the Turtle was always getting in trouble. Sometimes he'd reach into the freezer and empty out all the food. Other times he'd sled on the deck and get a splinter. His aunt Jane tried as hard as she could to keep him out of trouble, but he was sneaky and got into lots of trouble behind her back.One day, James thought he would go into town and see what kind of trouble he could get into. He went to the grocery store and pulled all the pudding off the shelves and ate two jars. Then he walked to the fast food restaurant and ordered 15 bags of fries. He didn't pay, and instead headed home.
His aunt was waiting for him in his room. She told James that she loved him, but he would have to start acting like a well-behaved turtle.After about a month, and after getting into lots of trouble, James finally made up his mind to be a better turtle.
1) What is the name of the trouble making turtle?
A) Fries
B) Pudding
C) James
D) Jane
2) What did James pull off of the shelves in the grocery store?
A) pudding
B) fries
C) food
D) splinters
3) Where did James go after he went to the grocery store?
A) his deck
B) his freezer
C) a fast food restaurant
D) his room
4) What did James do after he ordered the fries?
A) went to the grocery store
B) went home without paying
C) ate them
D) made up his mind to be a better turtle
-
In recent years many news outlets have begun to include summary points with each news article, presumably to suit the short attention spans of online readers. Notably, these summary points are not merely extractions of text contained in the article, but summary points which can be used to automatically construct queries which may require comprehension of the news article in order to answer.As shown the snippet below, the query is constructed by removing an entity from the statement, and tasking the reader to fill in the most appropriate entity from the text, in a manner similar to Cloze questions. Entities are identified and coreferenced in pre-processing, and are entirely masked in the text. This is intended to prevent the model from using external knowledge about the entities when choosing an answer, and instead rely on its understanding of the contexts in which the entities appear.
There are two datasets, one comprised of CNN news articles, the other of Daily Mail -- a UK-based tabloid. The CNN dataset has 90,266 documents for training, and 1,093 for test. With an average of four queries per document, this yields 380,298 queries for training and 3,198 for test. The Daily Mail section is significantly larger, having roughly twice as much training data and ten times as much test.
@entity0 , @entity1 ( @entity2 ) @entity3 lure @entity5 and @entity6 migrants by offering discounts to get onto overcrowded ships if people bring more potential passengers , a @entity2 investigation has revealed . a smuggler in the @entity1 capital of @entity0 laid bare the system for loading boats with poor and desperate refugees , during a conversation that a @entity2 producer secretly filmed . the conversation , recorded using a mobile phone , exposes the prices and incentives used to gather as many migrants as possible onto ships . an estimated 1,600 migrants have died so far this year on the dangerous @entity26 crossing , but still more wait to try to reach @entity27 . @entity2 's producer was introduced to a @entity30 and @entity31 smuggler by an intermediary in @entity0 , who mistakenly thought she was a @entity34 looking to bring other @entity34 refugees with her onto boats to @entity37 . why i fled : migrants share their stories the smuggler took her to an unfinished building on the outskirts of @entity0 near the city 's many ports , where the migrants they have already found are kept until the crossing is ready . the building could only be reached by walking down a trash - littered alleyway , and featured a series of packed rooms , separated by curtains , where dozens sat -- well over the 80 migrants she was promised would be in her boat . the smuggler explained that the " final price " for @entity34 -- often thought to be richer than their african migrant counterparts -- was $ 1,000 . he added that for each @entity34 she brought with her , the producer would get a $ 100 discount . so if she brought 10 , she could travel free . he explained how the " discount " was " well known , " suggesting perhaps it was part of the unwritten rules that govern the trade and why so many migrants come to each boat . any fears about the crossing were supposed to be allayed by the smuggler insisting the boats they used had new motors , and that the @entity30 pilot would have a satellite telephone and gps to assist the crossing . he also assured @entity2 's producer , when asked , that if the people became too many , they would use two boats . pregnant women among migrants trying to cross@placeholder investigation uncovers the business inside a human smuggling ring
-
Just as DeepMind used Cloze style transformations to turn summary points into queries, researchers at FAIR applied this technique to construct queries from the sentences which follow a block of text. In the Children's Book Test, a collection of children's books were gathered from the Project Gutenberg archives. Each question is constructed by taking 20 consecutive sentences from the book text, and leaving the 21st as the query statement. A word from the query is selected and masked, and the reader is tasked with selecting which word from the text (of the chosen type) should be used to fill this placeholder in the query.CBT differs from the DeepMind CNN and DailyMail datasets, as it is not merely entities that are tested. Named entities, common nouns, verbs, and prepositions are the four categories of words which may be treated as placeholders, and for each the complete set of candidates words is identified in the text using off-the-shelf NLP tools (Stanford Core NLP). It also differs in that entities in the text are not masked, and the model is encouraged to build up a persistent representation of the participants in the story as it reads through the book text.
` What is it ? '
answered he .
` The ogre is coming after us .
I saw him . '
` But where is he ?
I do n't see him . '
` Over there .
He only looks about as tall as a needle . '
Then they both began to run as fast as they could , while the ogre and his dog kept drawing always nearer .
A few more steps , and he would have been by their side , when Dschemila threw the darning needle behind her .
In a moment it became an iron mountain between them and their enemy .
` We will break it down , my dog and I , ' cried the ogre in a rage , and they dashed at the mountain till they had forced a path through , and came ever nearer and nearer .
` Cousin ! '
said Dschemila suddenly .
` What is it ? '
` The ogre is coming after us with his dog . '
` You go on in front then , ' answered he ; and they both ran on as fast as they could , while the ogre and the dog drew always nearer and nearer .
` They are close upon us ! '
cried the maiden , glancing behind , ` you must throw the pin . '
So Dschemil took the pin from his cloak and threw it behind him , and a dense thicket of thorns sprang up round them , which the ogre and his dog could not pass through . 'I will get through it somehow , if I burrow underground , ' cried he , and very soon he and the XXXXX were on the other side .
[gold]: dog
[answer canditates]: Cousin | cloak | dog | maiden | mountain | needle | path | pin | side | steps
-
The Stanford Question Answering Dataset consists of 107,785 question-answer pairs on 536 articles. The text passages are taken from Wikipedia across a wide range of topics, and the question-answer pairs themselves are human annotated via crowdsourcing.SQuAD is notable in that while the answers are contained verbatim within the corresponding text passage, they need not be entities and sets of candidates answers are not provided. This makes SQuAD the first large scale QA dataset where answers are spans of text, which must be identified without additional clues.
Conventionally, a computer consists of at least one processing element, typically a central processing unit (CPU), and some form of memory. The processing element carries out arithmetic and logic operations, and a sequencing and control unit can change the order of operations in response to stored information. Peripheral devices allow information to be retrieved from an external source, and the result of operations saved and retrieved.In computer terms, what does CPU stand for?
What are the devices called that are from an external source?
What are two things that a computer always has?