why you should care about byte level seq2seq models in nlp
play

Why You Should Care About Byte-Level Seq2Seq Models in NLP South - PowerPoint PPT Presentation

Why You Should Care About Byte-Level Seq2Seq Models in NLP South England Natural Language Processing Meetup Alan Turing Institute Monday March 4, 2019 Tom Kenter TTS Research Google UK, London Based on internship at Google Research in


  1. Why You Should Care About Byte-Level Seq2Seq Models in NLP South England Natural Language Processing Meetup Alan Turing Institute Monday March 4, 2019 Tom Kenter TTS Research Google UK, London

  2. Based on internship at Google Research in Mountain View Byte-level Machine Reading across Morphologically Varied Languages Tom Kenter, Llion Jones, Daniel Hewlett Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), 2018 https://ai.google/research/pubs/pub47437 Medium blogpost Why You Should Care About Byte-Level Sequence-to-Sequence Models in NLP Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  3. Proprietary + Confidential Is it advantageous, when processing morphologically rich languages, to use bytes rather than words as input and output to RNNs in a machine reading task?

  4. Machine Reading Computer reads text and has to answer questions about it. WikiReading datasets English WikiReading dataset ● (Hewlett, et al, ACL, 2016) Two extra datasets — Russian and Turkish — ● (Kenter et al, AAAI, 2018) https://github.com/google-research-datasets/wiki-reading Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  5. Byte-level Machine Reading s d 1 1 1 n 0 1 0 a l 0 1 1 r e 0 0 0 h 1 1 0 e t 1 0 0 h e n 1 1 1 N t I 1 0 1 word-level byte-level e s m r i e 1 1 1 a 1 1 0 h d W 1 0 1 r 0 0 1 e 0 1 0 t s 1 1 0 m 0 0 1 A 0 0 1 Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  6. Morphologically rich languages Russian Turkish В прошлом году Дмитрий переехал в kolay → easy kolayla ş tırabiliriz Москву. → we can make it easier kolayla ş tıramıyoruz → we cannot make it easier Где теперь живет Дмитрий? В Москве. Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  7. Why should you care about byte-level seq2seq models in NLP? 1 1 1 0 1 0 Small input vocabulary → small model size 0 1 1 0 0 0 1 1 0 1 0 0 1 1 1 1 0 1 No out-of-vocabulary problem Allows for apples-to-apples comparison between models Universal encoding scheme across languages byte-level Longer unroll length for RNN 1 1 1 1 1 0 1 0 1 0 0 1 ⟷ read less input 0 1 0 1 1 0 0 0 1 0 0 1 Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  8. Models Multi-level RNN Bidirectional RNN Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  9. Models Hybrid word-byte model Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  10. Models Convolutional recurrent Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  11. Models Memory network Encoder-transformer-decoder Memory network/encoder-transformer-decoder Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  12. Results Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  13. Results Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  14. Conclusions Reading and outputting bytes, instead of words, works. Byte-level models provide an elegant way of dealing with the out-of-vocabulary problem. Byte-level models perform on par with the state-of-the-art word-level model on English, and better on morphologically more involved languages. This is good news, as byte-level models have far fewer parameters. Source: Lorem ipsum dolor sit amet, consectetur adipiscing elit. Duis non erat sem

  15. Are you interested in machine reading/question answering/NLU, and looking for a new challenge? Try your approach on 3 languages at once! WikiReading English & Russian & Turkish https://github.com/google-research-datasets/wiki-reading

  16. Thank you Byte-level Machine Reading across Morphologically Varied Languages Tom Kenter, Llion Jones, Daniel Hewlett Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), 2018 https://ai.google/research/pubs/pub47437 Medium blogpost Why You Should Care About Byte-Level Sequence-to-Sequence Models in NLP

Recommend


More recommend