natural language processing for analyzing disaster
play

Natural Language Processing for Analyzing Disaster Recovery Trends - PowerPoint PPT Presentation

Natural Language Processing for Analyzing Disaster Recovery Trends Expressed in Large Text Corpora Lucy H. Lin , Scott B. Miles, Noah A. Smith University of Washington 19 October 2018 Introduction Problem : empirical data describing disaster


  1. Natural Language Processing for Analyzing Disaster Recovery Trends Expressed in Large Text Corpora Lucy H. Lin , Scott B. Miles, Noah A. Smith University of Washington 19 October 2018

  2. Introduction Problem : empirical data describing disaster recovery is scarce Available data source: text corpora (e.g., news articles) Manual analysis of large text corpora is slow… ⇒ constrains pre-/post-event recovery planning ⇒ + natural language processing

  3. Introduction anger among homeowners.” freq Proposition query: Frequency across time: aggregate “Those in charge of recovery are making moves to appease the growing “Unfamiliar bureaucratic systems are causing the majority of the stress.” Matched sentences: “Dealing with authorities is causing stress and anxiety.” query corpus 2 3 4 5 1 1 1 1 1 1 0 0 0 0 0 2 2 2 2 2

  4. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  5. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  6. 2010–2011 Canterbury earthquake disaster: timeline September 2010 : • Epicenter: 35km west of Christchurch • Moderate damage February 2011 : • Epicenter: 10km southeast of Christchurch • Extremely high ground acceleration • 185 deaths, thousands of felt aftershocks

  7. 2010–2011 Canterbury earthquake disaster: impacts Damages : • Estimated $40 billion • Housing: 100k houses in need of repairs • Water, utilities, road infrastructure: extensive damage Recovery groups : • Government: CERA, SCIRT (sunset after 5 years) • Community: Regenerate Christchurch Recovery still ongoing: public development projects, residential rezoning

  8. 2010–2011 Canterbury earthquake disaster: text data Corpus : 982 NZ news articles (2010–2015) post-earthquakes • stuff.co.nz , nzherald.co.nz • Community wellbeing • Infrastructure • Decision-making e.g.: “The council should have consulted residents before making decisions.” Proposition queries : 20 queries, covering

  9. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  10. Semantic matching Goal : find sentences with similar meaning to the query. • Needs to be more powerful than word/phrase-level matching. • Related to information retrieval, but want all matches .

  11. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  12. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  13. Semantic matching: fast filter . . stress average Goal : quickly filter out unlikely matches. Dealing with . . . . . . anxiety cosine . similarity . bureaucratic Word vector based comparison between two sentences: average . Unfamiliar .   ⊏ ⊐  ⊏ ⊐  corpus sent. ⊏ ⊐         ⊏ ⊐   ⊏ ⊐  ⊏ ⊐    query sent. ⊏ ⊐       ⊏ ⊐

  14. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  15. Semantic matching: syntax-based model Finer-grained matching : take word order/syntax into account. Intuition : transformation between sentences is indicative of their relationship.

  16. Semantic matching: syntax-based model causing +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) ? query root stress is unfamiliar authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  17. Semantic matching: syntax-based model root +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing are systems ? query stress unfamiliar causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  18. Semantic matching: syntax-based model root +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing are dealing ? query stress unfamiliar causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  19. Semantic matching: syntax-based model root +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing is dealing ? query stress unfamiliar causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  20. Semantic matching: syntax-based model query +insert(authorities) +relabel(are) +relabel(systems) +delete(bureaucratic) +delete(unfamiliar) root stress causing is authorities with dealing root unfamiliar stress causing is authorities with dealing candidate root stress causing are systems bureaucratic +insert(with)

  21. Semantic matching: method overview fast filter corpus of sentences proposition query likely matches syntax-based model matched sentences

  22. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  23. User evaluation Questions : • How good are the sentences matched by our method? • Do potential users think this kind of tool will be helpful? User study : 20 emergency managers

  24. User evaluation: output quality Rated output from 20 proposition queries: • Different method variants • Different parts of method: • Not selected by filter • Selected by filter, but not part of final output • Top-scoring output from filter • Method output (from syntax-based model) • 1-5 scale (Krippendorf’s α = 0 . 784)

  25. User evaluation: example Query: There is a shortage of construction workers. “The quarterly report for Canterbury included analysis on ( 1 : completely unrelated to the query) Greater Christchurch Value of Work projections.”

  26. User evaluation: example Query: There is a shortage of construction workers. December.” ( 3 : related to but does not adequately express the query) “The construction sector’s workload was expected to peak in

  27. User evaluation: example Query: There is a shortage of construction workers. and was likely to remain that way.” ( 5 : expresses the query in its entirety) “Greater Christchurch’s labour supply for the rebuild was tight

  28. User evaluation: results 1 2 3 Average score Best performing system Not selected by filter Selected by filter (unmatched) 3 . 22 3 . 1 2 . 5 2 . 03 1 . 5 1 . 06

  29. User evaluation: results 1 2 3 Average score Best performing system Not selected by filter Selected by filter (unmatched) Highest-scoring by filter 3 . 22 3 . 1 2 . 5 2 . 03 1 . 5 1 . 06

  30. User evaluation: results 1 2 3 Average score Best performing system Not selected by filter Selected by filter (unmatched) Highest-scoring by filter Matched by method 3 . 22 3 . 1 2 . 5 2 . 03 1 . 5 1 . 06

  31. User evaluation: is this interesting? Other feedback : news/other text corpora • 17/20 respondents interested in measuring ideas in

  32. User evaluation: round two Follow-up study : • 7 return participants • Replicated findings of first user study • Participant-supplied queries (18)

  33. Outline 1. Introduction 2. Case study: 2010-2011 Canterbury earthquake disaster 3. NLP method for semantic matching 4. User evaluation 5. Qualitative/quantitative output 6. Conclusions

  34. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly.

  35. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly. “Orion Energy CEO Roger Sutton says most of the west of Christchurch now has fully restored power.”

  36. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly. “He had no water but power had been restored in his area.”

  37. Recovery trends: example #1 2010 2011 2012 2013 2014 2015 0 3 6 9 12 Frequency The power system was fully restored quickly. “TV3 reports that power has now been restored to 60 per cent of Christchurch.”

  38. Recovery trends: example #1 6 restore power and the situation could remain for the next few “It had been unable to access the electricity network to The power system was fully restored quickly. Frequency 12 9 3 2010 0 2015 2014 2013 2012 2011 days.”

  39. Recovery trends: example #2 2010 2011 2012 2013 2014 2015 0 2 4 6 Frequency Dealing w/authorities is causing stress and anxiety.

Recommend


More recommend