deep learning for language understanding at google scale
play

Deep Learning for Language Understanding (at Google Scale) Anjuli - PowerPoint PPT Presentation

Deep Learning for Language Understanding (at Google Scale) Anjuli Kannan Software Engineer, Google Brain Confidential + Proprietary Confidential + Proprietary Text is just a sequence of words ["hi", "team",


  1. Deep Learning for Language Understanding (at Google Scale) Anjuli Kannan Software Engineer, Google Brain Confidential + Proprietary Confidential + Proprietary

  2. Text is just a sequence of words ["hi", "team", "the", "server", "appears", "to", "be", "dropping", "about", "10%", …] Confidential + Proprietary

  3. About me ● My team: Google Brain ○ "Make machines intelligent, improve people's lives." ○ Research + software + applications ○ g.co/brain My work is at boundary of research and applications ● ● Focus on natural language understanding

  4. Neural network basics

  5. Neural network ... Is a 4 Is a 5 ... Image: Wikipedia Confidential + Proprietary

  6. Neural network Is a 4 Is a 5 Neuron Confidential + Proprietary

  7. Basic building block is the neuron Greg Corrado Confidential + Proprietary

  8. Gradient descent Learning Rate w’ = w - α ∂ w L(w) w w’ Slide: Vincent Vanhoucke

  9. Recurrent neural networks

  10. Recurrent neural networks can model sequences Confidential + Proprietary

  11. Recurrent neural networks can model sequences How Message

  12. Recurrent neural networks can model sequences How are Message

  13. Recurrent neural networks can model sequences How are you Message

  14. Recurrent neural networks can model sequences How are you ? Message

  15. Recurrent neural networks can model sequences Internal state is a fixed length encoding of the message How are you ? Message

  16. Sequence-to-sequence models

  17. Suppose we want to generate email replies Response Incoming Smartreply email email

  18. Sequence-to-sequence model Sutskever et al, NIPS 2014

  19. Sequence-to-sequence model decoder encoder

  20. Sequence-to-sequence model Generates reply message Ingests incoming message

  21. Encoder ingests the incoming message Internal state is a fixed length encoding of the message How are you ? Message

  22. Decoder is initialized with final state of encoder How How are are you you ? ? __ Message

  23. Decoder is initialized with final state of encoder How How are are you you ? ? __ Message

  24. Decoder predicts next word Response I How are you ? __ Message

  25. Decoder predicts next word Response I am How are you ? __ I Message

  26. Decoder predicts next word Response I am great How are you ? __ I am Message

  27. Decoder predicts next word Response I am ! great How are you ? __ I am great Message Vinyals & Le, ICML DL 2015 Kannan et al, KDD 2016

  28. What the model can do

  29. What the model can do

  30. Summary - Neural networks learn feature representations from raw data - Recurrent neural networks have statefulness which allows them to model sequences of data such as text - The sequence-to-sequence model contains two recurrent neural networks: one to encode an input sequence and one to generate an output sequence

  31. Smartreply

  32. Google Translate

  33. Research: Speech recognition

  34. Research: Electronic health records

  35. What's next? ?

  36. Resources - All tensorflow tutorials: https://www.tensorflow.org/versions/master/tutorials/index.html - Sequence-to-sequence tutorial (machine translation): https://www.tensorflow.org/versions/master/tutorials/seq2seq - Chris Olah's blog: http://colah.github.io/

  37. Thank you!

Recommend


More recommend