discourse coreference
play

Discourse: Coreference Deep Processing Techniques for NLP Ling 571 - PowerPoint PPT Presentation

Discourse: Coreference Deep Processing Techniques for NLP Ling 571 March 5, 2014 Roadmap Coreference Referring expressions Syntactic & semantic constraints Syntactic & semantic preferences Reference


  1. Discourse: Coreference Deep Processing Techniques for NLP Ling 571 March 5, 2014

  2. Roadmap — Coreference — Referring expressions — Syntactic & semantic constraints — Syntactic & semantic preferences — Reference resolution: — Hobbs Algorithm: Baseline — Machine learning approaches — Sieve models — Challenges

  3. Reference and Model

  4. Reference Resolution — Queen Elizabeth set about transforming her husband, King George VI, into a viable monarch. Logue, a renowned speech therapist, was summoned to help the King overcome his speech impediment... Coreference resolution: Find all expressions referring to same entity, ‘corefer’ Colors indicate coreferent sets Pronominal anaphora resolution: Find antecedent for given pronoun

  5. Referring Expressions — Indefinite noun phrases (NPs): e.g. “ a cat ” — Introduces new item to discourse context — Definite NPs: e.g. “ the cat ” — Refers to item identifiable by hearer in context — By verbal, pointing, or environment availability; implicit — Pronouns: e.g. “ he ” , ” she ” , “ it ” — Refers to item, must be “ salient ” — Demonstratives: e.g. “ this ” , “ that ” — Refers to item, sense of distance (literal/figurative) — Names: e.g. “Miss Woodhouse”,”IBM” — New or old entities

  6. Information Status — Some expressions (e.g. indef NPs) introduce new info — Others refer to old referents (e.g. pronouns) — Theories link form of refexp to given/new status — Accessibility: — More salient elements easier to call up, can be shorter Correlates with length: more accessible, shorter refexp

  7. Complicating Factors — Inferrables: — Refexp refers to inferentially related entity — I bought a car today, but the door had a dent, and the engine was noisy. — E.g. car -> door, engine — Generics: — I want to buy a Mac. They are very stylish. — General group evoked by instance. — Non-referential cases: — It’s raining.

  8. Syntactic Constraints for Reference Resolution — Some fairly rigid rules constrain possible referents — Agreement: — Number: Singular/Plural — Person: 1st: I,we; 2nd: you; 3rd: he, she, it, they — Gender: he vs she vs it

  9. Syntactic & Semantic Constraints — Binding constraints: — Reflexive (x-self): corefers with subject of clause — Pronoun/Def. NP: can ’ t corefer with subject of clause — “ Selectional restrictions ” : — “ animate ” : The cows eat grass. — “ human ” : The author wrote the book. — More general: drive: John drives a car….

  10. Syntactic & Semantic Preferences — Recency: Closer entities are more salient — The doctor found an old map in the chest. Jim found an even older map on the shelf. It described an island. — Grammatical role: Saliency hierarchy of roles — e.g. Subj > Object > I. Obj. > Oblique > AdvP — Billy Bones went to the bar with Jim Hawkins. He called for a glass of rum. [he = Billy] — Jim Hawkins went to the bar with Billy Bones. He called for a glass of rum. [he = Jim]

  11. Syntactic & Semantic Preferences — Repeated reference: Pronouns more salient — Once focused, likely to continue to be focused — Billy Bones had been thinking of a glass of rum. He hobbled over to the bar. Jim Hawkins went with him. He called for a glass of rum. [he=Billy] — Parallelism: Prefer entity in same role — Silver went with Jim to the bar. Billy Bones went with him to the inn. [him = Jim] — Overrides grammatical role — Verb roles: “ implicit causality ” , thematic role match,... — John telephoned Bill. He lost the laptop. [He=John] — John criticized Bill. He lost the laptop. [He=Bill]

  12. Reference Resolution Approaches — Common features — “ Discourse Model ” — Referents evoked in discourse, available for reference — Structure indicating relative salience — Syntactic & Semantic Constraints — Syntactic & Semantic Preferences — Differences: — Which constraints/preferences? How combine? Rank?

  13. Hobbs ’ Resolution Algorithm — Requires: — Syntactic parser — Gender and number checker — Input: — Pronoun — Parse of current and previous sentences — Captures: — Preferences: Recency, grammatical role — Constraints: binding theory, gender, person, number

  14. Hobbs Algorithm — Intuition: — Start with target pronoun — Climb parse tree to S root — For each NP or S — Do breadth-first, left-to-right search of children — Restricted to left of target — For each NP , check agreement with target — Repeat on earlier sentences until matching NP found

  15. Hobbs Algorithm Detail — Begin at NP immediately dominating pronoun — Climb tree to NP or S: X=node, p = path — Traverse branches below X, and left of p: BF , LR — If find NP , propose as antecedent — If separated from X by NP or S — Loop: If X highest S in sentence, try previous sentences. — If X not highest S, climb to next NP or S: X = node — If X is NP , and p not through X’s nominal, propose X — Traverse branches below X, left of p: BF ,LR — Propose any NP — If X is S, traverse branches of X, right of p: BF , LR — Do not traverse NP or S; Propose any NP — Go to Loop

  16. Hobbs Example Lyn’s mom is a gardener. Craige likes her.

  17. Another Hobbs Example — The castle in Camelot remained the residence of the King until 536 when he moved it to London. — What is it ? — residence

  18. Another Hobbs Example Hobbs, 1978

  19. Hobbs Algorithm — Results: 88% accuracy ; 90+% intrasentential — On perfect, manually parsed sentences — Useful baseline for evaluating pronominal anaphora — Issues: — Parsing: — Not all languages have parsers — Parsers are not always accurate — Constraints/Preferences: — Captures: Binding theory, grammatical role, recency — But not: parallelism, repetition, verb semantics, selection

  20. Data-driven Reference Resolution — Prior approaches: Knowledge-based, hand-crafted — Data-driven machine learning approach — Coreference as classification, clustering, ranking problem — Mention-pair model: — For each pair NPi,NPj, do they corefer? — Cluster to form equivalence classes — Entity-mention model — For each pair NP k and cluster C j,, should the NP be in the cluster? — Ranking models — For each NP k , and all candidate antecedents, which highest?

  21. NP Coreference Examples — Link all NPs refer to same entity Queen Elizabeth set about transforming her husband, King George VI, into a viable monarch. Logue, a renowned speech therapist, was summoned to help the King overcome his speech impediment... Example from Cardie&Ng 2004

  22. Annotated Corpora — Available shared task corpora — MUC-6, MUC-7 (Message Understanding Conference) — 60 documents each, newswire, English — ACE (Automatic Content Extraction) — Originally English newswite — Later include Chinese, Arabic; blog, CTS, usenet, etc — Treebanks — English Penn Treebank (Ontonotes) — German, Czech, Japanese, Spanish, Catalan, Medline

  23. Feature Engineering — Other coreference (not pronominal) features — String-matching features: — Mrs. Clinton <->Clinton — Semantic features: — Can candidate appear in same role w/same verb? — WordNet similarity — Wikipedia: broader coverage — Lexico-syntactic patterns: — E.g. X is a Y

  24. Typical Feature Set — 25 features per instance: 2NPs, features, class — lexical (3) — string matching for pronouns, proper names, common nouns — grammatical (18) — pronoun_1, pronoun_2, demonstrative_2, indefinite_2, … — number, gender, animacy — appositive, predicate nominative — binding constraints, simple contra-indexing constraints, … — span, maximalnp, … — semantic (2) — same WordNet class — alias — positional (1) — distance between the NPs in terms of # of sentences — knowledge-based (1) — naïve pronoun resolution algorithm

  25. Coreference Evaluation — Key issues: — Which NPs are evaluated? — Gold standard tagged or — Automatically extracted — How good is the partition? — Any cluster-based evaluation could be used (e.g. Kappa) — MUC scorer: — Link-based: ignores singletons; penalizes large clusters — Other measures compensate

  26. Clustering by Classification — Mention-pair style system: — For each pair of NPs, classify +/- coreferent — Any classifier — Linked pairs form coreferential chains — Process candidate pairs from End to Start — All mentions of an entity appear in single chain — F-measure: MUC-6: 62-66%; MUC-7: 60-61% — Soon et. al, Cardie and Ng (2002)

  27. Multi-pass Sieve Approach — Raghunathan et al., 2010 — Key Issues: — Limitations of mention-pair classifier approach — Local decisions over large number of features — Not really transitive — Can’t exploit global constraints — Low precision features may overwhelm less frequent, high precision ones

Recommend


More recommend