lexical semantics
stringclasses 11
values | predicate-argument structure
stringclasses 24
values | logic
stringclasses 30
values | knowledge
stringclasses 3
values | domain
stringclasses 5
values | premise
stringlengths 11
296
| hypothesis
stringlengths 11
296
| label
int64 0
2
|
---|---|---|---|---|---|---|---|
Conditionals | Wikipedia | If Charles' left wing, commanded by Nauendorf, united with Hotze's force, approaching from the east, Masséna knew Charles would attack and very likely push him out of Zürich. | If Charles' left wing, commanded by Nauendorf, united with Hotze's force, approaching from the east, Masséna would prepare for Charles to attack and very likely push him out of Zürich. | 1 |
|||
Conditionals | Wikipedia | If Charles' left wing, commanded by Nauendorf, united with Hotze's force, approaching from the east, Masséna would prepare for Charles to attack and very likely push him out of Zürich. | If Charles' left wing, commanded by Nauendorf, united with Hotze's force, approaching from the east, Masséna knew Charles would attack and very likely push him out of Zürich. | 1 |
|||
Datives | Wikipedia | Ferdinand of Naples refused to pay agreed-upon tribute to France, and his subjects followed this refusal with a rebellion. | Ferdinand of Naples refused to pay France the agreed-upon tribute, and his subjects followed this refusal with a rebellion. | 0 |
|||
Datives | Wikipedia | Ferdinand of Naples refused to pay France the agreed-upon tribute, and his subjects followed this refusal with a rebellion. | Ferdinand of Naples refused to pay agreed-upon tribute to France, and his subjects followed this refusal with a rebellion. | 0 |
|||
Coordination scope | Wikipedia | Furthermore, the French dangerously underestimated Austrian tenacity and military skill. | Furthermore, the French dangerously underestimated Austrian military skill and tenacity. | 0 |
|||
Coordination scope | Wikipedia | Furthermore, the French dangerously underestimated Austrian military skill and tenacity. | Furthermore, the French dangerously underestimated Austrian tenacity and military skill. | 0 |
|||
Coordination scope | Wikipedia | Furthermore, the French dangerously underestimated Austrian tenacity and military skill. | Furthermore, the French dangerously underestimated Austrian military skill. | 0 |
|||
Coordination scope | Wikipedia | Furthermore, the French dangerously underestimated Austrian military skill. | Furthermore, the French dangerously underestimated Austrian tenacity and military skill. | 1 |
|||
Quantifiers | Wikipedia | There are four supraocular scales (above the eyes) in almost all specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | There are four supraocular scales (above the eyes) in most specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | 0 |
|||
Quantifiers | Wikipedia | There are four supraocular scales (above the eyes) in most specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | There are four supraocular scales (above the eyes) in almost all specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | 1 |
|||
Redundancy | Restrictivity | Non-monotone | Wikipedia | There are four supraocular scales (above the eyes) in almost all specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | There are four scales above the eyes in almost all specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | 0 |
|
Redundancy | Restrictivity | Non-monotone | Wikipedia | There are four scales above the eyes in almost all specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | There are four supraocular scales (above the eyes) in almost all specimens and five supraciliary scales (immediately above the eyes, below the supraoculars). | 0 |
|
Universal | Wikipedia | All 860 officers and men on board, including Spee, went down with the ship. | Spee went down with the ship. | 0 |
|||
Universal | Wikipedia | Spee went down with the ship. | All 860 officers and men on board, including Spee, went down with the ship. | 1 |
|||
Lexical entailment | Downward monotone;Existential;Negation | Wikipedia | Regional governors could not rely on the king for help in times of crisis, and the ensuing food shortages and political disputes escalated into famines and small-scale civil wars. | Regional governors could not rely on anyone for help in times of crisis, and the ensuing food shortages and political disputes escalated into famines and small-scale civil wars. | 1 |
||
Lexical entailment | Downward monotone;Existential;Negation | Wikipedia | Regional governors could not rely on anyone for help in times of crisis, and the ensuing food shortages and political disputes escalated into famines and small-scale civil wars. | Regional governors could not rely on the king for help in times of crisis, and the ensuing food shortages and political disputes escalated into famines and small-scale civil wars. | 0 |
||
World knowledge | Wikipedia | The pharaohs of the Middle Kingdom restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | The pharaohs of the Middle Kingdom of Egypt restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | 0 |
|||
World knowledge | Wikipedia | The pharaohs of the Middle Kingdom of Egypt restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | The pharaohs of the Middle Kingdom restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | 0 |
|||
World knowledge | Wikipedia | The pharaohs of the Middle Kingdom restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | The pharaohs of the Middle Kingdom of China restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | 2 |
|||
World knowledge | Wikipedia | The pharaohs of the Middle Kingdom of China restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | The pharaohs of the Middle Kingdom restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | 2 |
|||
World knowledge | Wikipedia | The pharaohs of the Middle Kingdom restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | The pharaohs of Egypt restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | 0 |
|||
World knowledge | Wikipedia | The pharaohs of Egypt restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | The pharaohs of the Middle Kingdom restored the country's stability and prosperity, thereby stimulating a resurgence of art, literature, and monumental building projects. | 1 |
|||
Redundancy | Upward monotone | Wikipedia | The 15th Tank Corps was a tank corps of the Soviet Union's Red Army. | The 15th Tank Corps was a corps of the Soviet Union's Red Army. | 0 |
||
Redundancy | Upward monotone | Wikipedia | The 15th Tank Corps was a corps of the Soviet Union's Red Army. | The 15th Tank Corps was a tank corps of the Soviet Union's Red Army. | 0 |
||
Factivity | Negation | Artificial | I can't believe it's not butter. | It's not butter. | 1 |
||
Factivity | Negation | Artificial | It's not butter. | I can't believe it's not butter. | 1 |
||
Factivity | Negation | Artificial | I can't believe it's not butter. | It's butter. | 1 |
||
Factivity | Negation | Artificial | It's butter. | I can't believe it's not butter. | 1 |
||
Quantifiers | ACL | However, these regularities are sometimes obscured by semantic and syntactic differences. | However, these regularities are always obscured by semantic and syntactic differences. | 1 |
|||
Quantifiers | ACL | However, these regularities are always obscured by semantic and syntactic differences. | However, these regularities are sometimes obscured by semantic and syntactic differences. | 1 |
|||
Quantifiers | Conjunction | ACL | However, these regularities are sometimes obscured by semantic and syntactic differences. | However, these regularities are sometimes obscured by syntactic differences. | 0 |
||
Quantifiers | Conjunction | ACL | However, these regularities are sometimes obscured by syntactic differences. | However, these regularities are sometimes obscured by semantic and syntactic differences. | 1 |
||
Redundancy | Intersectivity | Upward monotone | World knowledge | ACL | In grounded communication tasks, speakers face pressures in choosing referential expressions that distinguish their targets from others in the context, leading to many kinds of pragmatic meaning enrichment. | In grounded communication tasks, speakers face pressures in choosing referential expressions that distinguish their targets from others in the context, leading to many kinds of meaning enrichment. | 0 |
Redundancy | Intersectivity | Upward monotone | World knowledge | ACL | In grounded communication tasks, speakers face pressures in choosing referential expressions that distinguish their targets from others in the context, leading to many kinds of meaning enrichment. | In grounded communication tasks, speakers face pressures in choosing referential expressions that distinguish their targets from others in the context, leading to many kinds of pragmatic meaning enrichment. | 0 |
Ellipsis/Implicits | ACL | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the others. | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the other colors. | 0 |
|||
Ellipsis/Implicits | ACL | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the other colors. | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the others. | 0 |
|||
Ellipsis/Implicits | ACL | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the others. | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the other utterances. | 2 |
|||
Ellipsis/Implicits | ACL | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the other utterances. | Thus, a model of the speaker must process representations of the colors in the context and produce an utterance to distinguish the target color from the others. | 2 |
|||
Restrictivity | Non-monotone | ACL | While most successful approaches for reading comprehension rely on recurrent neural networks (RNNs), running them over long documents is prohibitively slow because it is difficult to parallelize over sequences. | While most approaches for reading comprehension rely on recurrent neural networks (RNNs), running them over long documents is prohibitively slow because it is difficult to parallelize over sequences. | 1 |
||
Restrictivity | Non-monotone | ACL | While most approaches for reading comprehension rely on recurrent neural networks (RNNs), running them over long documents is prohibitively slow because it is difficult to parallelize over sequences. | While most successful approaches for reading comprehension rely on recurrent neural networks (RNNs), running them over long documents is prohibitively slow because it is difficult to parallelize over sequences. | 1 |
||
Quantifiers | ACL | Due to the structure and short length of most Wikipedia documents (median number of sentences: 9), the answer can usually be inferred from the first few sentences. | Due to the structure and short length of most Wikipedia documents (median number of sentences: 9), the answer can always be inferred from the first few sentences. | 1 |
|||
Quantifiers | ACL | Due to the structure and short length of most Wikipedia documents (median number of sentences: 9), the answer can always be inferred from the first few sentences. | Due to the structure and short length of most Wikipedia documents (median number of sentences: 9), the answer can usually be inferred from the first few sentences. | 0 |
|||
Universal;Conjunction | ACL | Each captures only a single aspect of coherence, and all focus on scoring existing sentences, rather than on generating coherent discourse for tasks like abstractive summarization. | Each captures only a single aspect of coherence and focuses on scoring existing sentences, rather than on generating coherent discourse for tasks like abstractive summarization. | 0 |
|||
Universal;Conjunction | ACL | Each captures only a single aspect of coherence and focuses on scoring existing sentences, rather than on generating coherent discourse for tasks like abstractive summarization. | Each captures only a single aspect of coherence, and all focus on scoring existing sentences, rather than on generating coherent discourse for tasks like abstractive summarization. | 0 |
|||
Factivity | ACL | In a coherent context, a machine should be able to guess the next utterance given the preceding ones. | In a coherent context, a machine can guess the next utterance given the preceding ones. | 1 |
|||
Factivity | ACL | In a coherent context, a machine can guess the next utterance given the preceding ones. | In a coherent context, a machine should be able to guess the next utterance given the preceding ones. | 1 |
|||
Relative clauses | ACL | We thus propose eliminating the influence of the language model, which yields the following coherence score. | The language model yields the following coherence score. | 2 |
|||
Relative clauses | ACL | The language model yields the following coherence score. | We thus propose eliminating the influence of the language model, which yields the following coherence score. | 2 |
|||
Relative clauses | ACL | We thus propose eliminating the influence of the language model, which yields the following coherence score. | Eliminating the influence of the language model yields the following coherence score. | 0 |
|||
Relative clauses | ACL | Eliminating the influence of the language model yields the following coherence score. | We thus propose eliminating the influence of the language model, which yields the following coherence score. | 1 |
|||
Factivity | ACL | The topic for the current sentence is drawn based on the topic of the preceding sentence (or word) rather than on the global document-level topic distribution in vanilla LDA. | The topic for the current sentence is drawn based on the global document-level topic distribution in vanilla LDA. | 2 |
|||
Factivity | ACL | The topic for the current sentence is drawn based on the global document-level topic distribution in vanilla LDA. | The topic for the current sentence is drawn based on the topic of the preceding sentence (or word) rather than on the global document-level topic distribution in vanilla LDA. | 2 |
|||
Factivity | ACL | The topic for the current sentence is drawn based on the topic of the preceding sentence (or word) rather than on the global document-level topic distribution in vanilla LDA. | The topic for the current sentence is drawn based on the topic of the preceding sentence (or word). | 0 |
|||
Factivity | ACL | The topic for the current sentence is drawn based on the topic of the preceding sentence (or word). | The topic for the current sentence is drawn based on the topic of the preceding sentence (or word) rather than on the global document-level topic distribution in vanilla LDA. | 0 |
|||
Coordination scope;Prepositional phrases | ACL | We publicly share our dataset and code for future research. | We publicly share our dataset for future research. | 0 |
|||
Coordination scope;Prepositional phrases | ACL | We publicly share our dataset for future research. | We publicly share our dataset and code for future research. | 1 |
|||
Coordination scope;Prepositional phrases | ACL | We publicly share our dataset and code for future research. | We code for future research. | 1 |
|||
Coordination scope;Prepositional phrases | ACL | We code for future research. | We publicly share our dataset and code for future research. | 1 |
|||
Datives | ACL | This gives the model a sense of the implied action dynamics of the verb between the agent and the world. | This gives to the model a sense of the implied action dynamics of the verb between the agent and the world. | 0 |
|||
Datives | ACL | This gives to the model a sense of the implied action dynamics of the verb between the agent and the world. | This gives the model a sense of the implied action dynamics of the verb between the agent and the world. | 0 |
|||
Datives | ACL | This gives the model a sense of the implied action dynamics of the verb between the agent and the world. | This gives the model to a sense of the implied action dynamics of the verb between the agent and the world. | 1 |
|||
Datives | ACL | This gives the model to a sense of the implied action dynamics of the verb between the agent and the world. | This gives the model a sense of the implied action dynamics of the verb between the agent and the world. | 1 |
|||
Lexical entailment | ACL | This attribute group specifies prominent body parts involved in carrying out the action. | This attribute group specifies prominent limbs involved in carrying out the action. | 0 |
|||
Lexical entailment | ACL | This attribute group specifies prominent limbs involved in carrying out the action. | This attribute group specifies prominent body parts involved in carrying out the action. | 1 |
|||
Temporal | ACL | This problem has been studied before for zero-shot object recognition, but there are several key differences. | This problem has been previously studied for zero-shot object recognition, but there are several key differences. | 0 |
|||
Temporal | ACL | This problem has been previously studied for zero-shot object recognition, but there are several key differences. | This problem has been studied before for zero-shot object recognition, but there are several key differences. | 0 |
|||
Temporal | ACL | This problem has been studied before for zero-shot object recognition, but there are several key differences. | This problem will be studied for zero-shot object recognition, but there are several key differences. | 1 |
|||
Temporal | ACL | This problem will be studied for zero-shot object recognition, but there are several key differences. | This problem has been studied before for zero-shot object recognition, but there are several key differences. | 1 |
|||
Coordination scope | ACL | Understanding a long document requires tracking how entities are introduced and evolve over time. | Understanding a long document requires evolving over time. | 1 |
|||
Coordination scope | ACL | Understanding a long document requires evolving over time. | Understanding a long document requires tracking how entities are introduced and evolve over time. | 1 |
|||
Coordination scope | ACL | Understanding a long document requires tracking how entities are introduced and evolve over time. | Understanding a long document requires tracking how entities evolve over time. | 0 |
|||
Coordination scope | ACL | Understanding a long document requires tracking how entities evolve over time. | Understanding a long document requires tracking how entities are introduced and evolve over time. | 1 |
|||
Conjunction;Upward monotone | ACL | Understanding a long document requires tracking how entities are introduced and evolve over time. | Understanding a long document requires understanding how entities are introduced. | 0 |
|||
Conjunction;Upward monotone | ACL | Understanding a long document requires understanding how entities are introduced. | Understanding a long document requires tracking how entities are introduced and evolve over time. | 1 |
|||
Factivity | ACL | We do not assume that these variables are observed at test time. | These variables are not observed at test time. | 1 |
|||
Factivity | ACL | These variables are not observed at test time. | We do not assume that these variables are observed at test time. | 1 |
|||
Double negation | ACL | To compute the perplexity numbers on the test data, our model only takes account of log probabilities on word prediction. | To compute the perplexity numbers on the test data, our model doesn't take account of anything other than the log probabilities on word prediction. | 0 |
|||
Double negation | ACL | To compute the perplexity numbers on the test data, our model doesn't take account of anything other than the log probabilities on word prediction. | To compute the perplexity numbers on the test data, our model only takes account of log probabilities on word prediction. | 0 |
|||
Disjunction | ACL | We also experiment with the option to either use the pretrained GloVe word embeddings or randomly initialized word embeddings (then updated during training). | We experiment with the option using randomly initialized word embeddings (then updated during training). | 0 |
|||
Disjunction | ACL | We experiment with the option using randomly initialized word embeddings (then updated during training). | We also experiment with the option to either use the pretrained GloVe word embeddings or randomly initialized word embeddings (then updated during training). | 1 |
|||
Disjunction;Non-monotone | ACL | The entity prediction task requires predicting xxxx given the preceding text either by choosing a previously mentioned entity or deciding that this is a “new entity”. | The entity prediction task requires predicting xxxx given the preceding text by choosing a previously mentioned entity. | 2 |
|||
Disjunction;Non-monotone | ACL | The entity prediction task requires predicting xxxx given the preceding text by choosing a previously mentioned entity. | The entity prediction task requires predicting xxxx given the preceding text either by choosing a previously mentioned entity or deciding that this is a “new entity”. | 2 |
|||
Intersectivity | Downward monotone | ACL | So there is no dedicated memory block for every entity and no distinction between entity mentions and non-mention words. | So there is no dedicated high-dimensional memory block for every entity and no distinction between entity mentions and non-mention words. | 0 |
||
Intersectivity | Downward monotone | ACL | So there is no dedicated high-dimensional memory block for every entity and no distinction between entity mentions and non-mention words. | So there is no dedicated memory block for every entity and no distinction between entity mentions and non-mention words. | 1 |
||
Existential | ACL | Our approach complements these previous methods. | Our approach complements some previous methods. | 0 |
|||
Existential | ACL | Our approach complements some previous methods. | Our approach complements these previous methods. | 1 |
|||
Intervals/Numbers | ACL | We manually annotated 687 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | We manually annotated over 650 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | 0 |
|||
Intervals/Numbers | ACL | We manually annotated over 650 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | We manually annotated 687 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | 1 |
|||
Intervals/Numbers | ACL | We manually annotated 687 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | We manually annotated over 690 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | 2 |
|||
Intervals/Numbers | ACL | We manually annotated over 690 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | We manually annotated 687 templates mapping KB predicates to text for different compositionality types (with 462 unique KB predicates), and use those templates to modify the original WebQuestionsSP question according to the meaning of the generated SPARQL query. | 2 |
|||
Relative clauses;Restrictivity | Conditionals | ACL | To generate diversity, workers got a bonus if the edit distance of a paraphrase was high compared to the MG question. | To generate diversity, workers whose paraphrases had high edit distance compared to the MG question got a bonus. | 0 |
||
Relative clauses;Restrictivity | Conditionals | ACL | To generate diversity, workers whose paraphrases had high edit distance compared to the MG question got a bonus. | To generate diversity, workers got a bonus if the edit distance of a paraphrase was high compared to the MG question. | 0 |
||
Relative clauses;Restrictivity | Conditionals | ACL | To generate diversity, workers got a bonus if the edit distance of a paraphrase was high compared to the MG question. | To generate diversity, workers got a bonus if the edit distance of a paraphrase was above 3 operations compared to the MG question. | 1 |
||
Relative clauses;Restrictivity | Conditionals | ACL | To generate diversity, workers got a bonus if the edit distance of a paraphrase was above 3 operations compared to the MG question. | To generate diversity, workers got a bonus if the edit distance of a paraphrase was high compared to the MG question. | 1 |
||
Lexical entailment | ACL | To generate complex questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | To generate simple questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | 2 |
|||
Lexical entailment | ACL | To generate simple questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | To generate complex questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | 2 |
|||
World knowledge | ACL | To generate complex questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | To generate highly compositional questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | 0 |
|||
World knowledge | ACL | To generate highly compositional questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | To generate complex questions we use the dataset WEBQUESTIONSSP, which contains 4,737 questions paired with SPARQL queries for Freebase. | 0 |
|||
Restrictivity | World knowledge | ACL | In this paper, we explore the idea of polyglot semantic translation, or learning semantic parsing models that are trained on multiple datasets and natural languages. | In this paper, we explore the idea of learning semantic parsing models that are trained on multiple datasets and natural languages. | 0 |
||
Restrictivity | World knowledge | ACL | In this paper, we explore the idea of learning semantic parsing models that are trained on multiple datasets and natural languages. | In this paper, we explore the idea of polyglot semantic translation, or learning semantic parsing models that are trained on multiple datasets and natural languages. | 0 |