strongly incremental repair detection
play

Strongly Incremental Repair Detection Julian Hough 1 , 2 and Matthew - PowerPoint PPT Presentation

Strongly Incremental Repair Detection Julian Hough 1 , 2 and Matthew Purver 2 1 Dialogue Systems Group and CITEC, University of Bielefeld 2 Cognitive Science Research Group, Queen Mary University of London October 26th 2014, EMNLP Doha, Qatar


  1. Strongly Incremental Repair Detection Julian Hough 1 , 2 and Matthew Purver 2 1 Dialogue Systems Group and CITEC, University of Bielefeld 2 Cognitive Science Research Group, Queen Mary University of London October 26th 2014, EMNLP Doha, Qatar Hough and Purver EMNLP 2014

  2. Problem statement 1 STIR: Strongly Incremental Repair Detection 2 Edit terms Repair start Reparandum start Repair end Evaluation measures for repair 3 Experiments and results 4 Conclusions and Future 5 Hough and Purver EMNLP 2014

  3. Problem statement 1 STIR: Strongly Incremental Repair Detection 2 Edit terms Repair start Reparandum start Repair end Evaluation measures for repair 3 Experiments and results 4 Conclusions and Future 5 Hough and Purver EMNLP 2014

  4. Self-repairs “But one of the, the two things that I’m really . . . ” “Our situation is just a little bit, kind of the opposite of that” “and you know it’s like you’re, I mean, employments are contractual by nature anyway” [Switchboard examples] Hough and Purver EMNLP 2014

  5. Self-repairs: Annotation scheme John [ likes + { uh } loves ] Mary � �� � ���� � �� � interregnum repair reparandum [Shriberg, 1994, onwards] Terminology: edit terms , interruption point (+), repair onset Hough and Purver EMNLP 2014

  6. Self-repairs: classes “But one of [ the, + the ] two things that I’m really . . . ” [repeat] “Our situation is just [ a little bit, + kind of the opposite ] of that” [substitution] “and you know it’s like [ you’re + { I mean } ] employments are contractual by nature anyway” [delete] [Switchboard examples] Hough and Purver EMNLP 2014

  7. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  8. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  9. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  10. Self-repair detection: why do we care? Dialogue systems (parsing speech) Hough and Purver EMNLP 2014

  11. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Hough and Purver EMNLP 2014

  12. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Self-repairs have meaning! Dialogue systems should not filter out the reparandum! Hough and Purver EMNLP 2014

  13. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Self-repairs have meaning! Dialogue systems should not filter out the reparandum! Accuracy evaluation Standard evaluation F-score on reparandum words Also interested in repair structure assignment! Hough and Purver EMNLP 2014

  14. Self-repair detection: why do we care? Interpreting self-repair Preserving the reparandum and repair structure Evidence: [Brennan and Schober, 2001] showed subjects use the reparandum to make faster decisions: “Pick the yell-purple square” faster “Pick the uhh-purple square” Self-repairs have meaning! Dialogue systems should not filter out the reparandum! Accuracy evaluation Standard evaluation F-score on reparandum words Also interested in repair structure assignment! Hough and Purver EMNLP 2014

  15. Self-repair detection: Incrementality Non-incremental vs. Incremental Dialogue Systems [Schlangen and Skantze, 2011] Hough and Purver EMNLP 2014

  16. Self-repair detection: Incrementality We want good incremental performance : Timing - Low latency, short time to detect repairs Evolution over time - Responsiveness of the detection (incremental accuracy) - Stability of the output (low jitter) Computational complexity - Minimal processing overhead (fast) Hough and Purver EMNLP 2014

  17. Self-repair detection Problem statement A system that achieves: Interpretation of repair - repair structure tags rather than just reparandum words Strong incrementality - Give the best results possible as early as possible - Computationally fast Controllable trade-off between incrementality and overall accuracy Hough and Purver EMNLP 2014

  18. Previous approaches: Noisy channel model Best coverage generative model [Zwarts et al., 2010, Johnson and Charniak, 2004] S-TAG exploits ( ‘rough copy’ ) dependency with string alignment [Zwarts et al., 2010] utterance-final F-score = 0.778 Hough and Purver EMNLP 2014

  19. Previous approaches: Noisy channel model Best coverage generative model [Zwarts et al., 2010, Johnson and Charniak, 2004] S-TAG exploits ( ‘rough copy’ ) dependency with string alignment [Zwarts et al., 2010] utterance-final F-score = 0.778 Two incremental measures: - Time-to-detection : 7.5 words from reparandum onset - 4.6 words from repair onset - Delayed accuracy : slow rise up to 6 words back Complexity O ( n 5 ) Hough and Purver EMNLP 2014

  20. Previous approaches: Noisy channel model Why poor incremental performance? Hough and Purver EMNLP 2014

  21. Previous approaches: Noisy channel model Why poor incremental performance? - Inherently non-incremental string-alignment - Utterance global (c.f. spelling correction) - Sparsity of alignment forms [Hough and Purver, 2013] Hough and Purver EMNLP 2014

  22. SOLUTION: Information theory and strong incrementality Local measures of fluency for minimum latency in detection Does not just rely on string alignment Information theoretic measures of language models [Keller, 2004, Jaeger and Tily, 2011] Minimal complexity Hough and Purver EMNLP 2014

  23. Problem statement 1 STIR: Strongly Incremental Repair Detection 2 Edit terms Repair start Reparandum start Repair end Evaluation measures for repair 3 Experiments and results 4 Conclusions and Future 5 Hough and Purver EMNLP 2014

  24. STIR: Strongly Incremental Repair Detection John [ likes + { uh } loves ] Mary � �� � ���� � �� � reparandum interregnum repair ... [ rm start ... rm end + { ed } rp start ... rp end ] ... Hough and Purver EMNLP 2014

  25. STIR: Strongly Incremental Repair Detection ... [ rm start ... rm end + { ed } rp start ... rp end ] ... ... { ed } ... Hough and Purver EMNLP 2014

  26. STIR: Strongly Incremental Repair Detection Hough and Purver EMNLP 2014

  27. STIR: Strongly Incremental Repair Detection “John” S 0 S 1 Hough and Purver EMNLP 2014

  28. STIR: Strongly Incremental Repair Detection “John” “likes” S 0 S 1 S 2 Hough and Purver EMNLP 2014

  29. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” ed S 0 S 1 S 2 S 3 ed Hough and Purver EMNLP 2014

  30. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start ? ed S 0 S 1 S 2 S 3 S 4 rp start ed Hough and Purver EMNLP 2014

  31. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start ? S 0 S 1 S 4 rm end ed S 2 S 3 rp start rm end ed Hough and Purver EMNLP 2014

  32. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start S 0 S 1 S 4 rm start rm end ed S 2 S 3 rp start rm start rm end ed Hough and Purver EMNLP 2014

  33. STIR: Strongly Incremental Repair Detection “John” “likes” “uh” “loves” rp start S 0 S 1 S 4 rm start rp sub end rm end ed S 2 S 3 rp start rp sub rm start rm end ed end Hough and Purver EMNLP 2014

  34. STIR: Strongly Incremental Repair Detection “Mary” “John” “likes” “uh” “loves” rp start S 0 S 1 S 4 S 5 rm start rp sub end rm end ed S 2 S 3 rp start rp sub rm start rm end ed end Hough and Purver EMNLP 2014

  35. STIR: fluency modelling using enriched n-gram LMs s ( w i − 2 , w i − 1 , w i ) (surprisal) WML( w i − 2 , w i − 1 , w i ) (syntactic fluency) H ( θ ( w | c )) (entropy) KL ( θ ( w | c a ) , θ ( w | c b )) (distribution divergence) Hough and Purver EMNLP 2014

  36. STIR: fluency modelling using enriched n-gram LMs s ( w i − 2 , w i − 1 , w i ) (surprisal) WML( w i − 2 , w i − 1 , w i ) (syntactic fluency) H ( θ ( w | c )) (entropy) KL ( θ ( w | c a ) , θ ( w | c b )) (distribution divergence) p lex (word) and p pos (POS) models Does not use lexical or POS values , but information theoretic measures [Keller, 2004, Jaeger and Tily, 2011, Clark et al., 2013] Hough and Purver EMNLP 2014

  37. STIR: fluency modelling using enriched n-gram LMs rp start local deviation from fluency : drop in WML lex 0.0 −0.2 −0.4 −0.6 WML −0.8 −1.0 −1.2 −1.4 any very care i havent had good really good experience with child Hough and Purver EMNLP 2014

  38. STIR: fluency modelling using enriched n-gram LMs Extend ‘rough copy’ dependency [Johnson and Charniak, 2004] to gradient measures Information content = entropy Parallelism = distributional similarity Hough and Purver EMNLP 2014

Recommend


More recommend