breaking nli systems
play

Breaking NLI Systems with Sentences that Require Simple Lexical - PowerPoint PPT Presentation

Breaking NLI Systems with Sentences that Require Simple Lexical Inferences Max Glockner 1 , Vered Shwartz 2 and Yoav Goldberg 2 1 TU Darmstadt 2 Bar-Ilan University July 18, 2018 SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural


  1. Breaking NLI Systems with Sentences that Require Simple Lexical Inferences Max Glockner 1 , Vered Shwartz 2 and Yoav Goldberg 2 1 TU Darmstadt 2 Bar-Ilan University July 18, 2018

  2. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  3. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  4. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  5. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  6. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  7. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street ⇒ NEUTRAL NEUTRAL Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  8. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street ⇒ NEUTRAL NEUTRAL 3. A magician performing for an audience in a nightclub Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  9. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street ⇒ NEUTRAL NEUTRAL 3. A magician performing for an audience in a nightclub ⇒ CONTRADICTION CONTRADICTION Event co-reference assumption Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  10. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Classifier Extract Features Premise Hypothesis Encoder Encoder Hypothesis Premise 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  11. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  12. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  13. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet SOTA exceeds human performance... 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  14. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet SOTA exceeds human performance... 1 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  15. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet SOTA exceeds human performance... 1 1 [Gururangan et al., 2018, Poliak et al., 2018]: by learning “easy clues” Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  16. Do neural NLI models implicitly learn lexical semantic relations? Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 4 / 13

  17. New Test Set We constructed a new test set to answer this question Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  18. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  19. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  20. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ w ′ is in the SNLI vocabulary and in pre-trained embeddings Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  21. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ w ′ is in the SNLI vocabulary and in pre-trained embeddings Crowdsourcing labels (mostly contradictions!) Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  22. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ w ′ is in the SNLI vocabulary and in pre-trained embeddings Crowdsourcing labels (mostly contradictions!) Contradiction The man is holding a saxophone → The man is holding an electric guitar Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

Recommend


More recommend