Sunday 15:15–16:00 in Small Room

Using deep learning in natural language processing: explaining Google's Neural Machine Translation

Rob Romijnders, Ahmet Taspinar

Audience level:
Intermediate

Description

Recent advancements in Natural Language Processing (NLP) use deep learning to improve performance. In September 2016, Google announced that Google Translate will shift from phrase-based to neural machine translation. Other fields of NLP are making a similar shift. This talk motivates and explains these algorithms and discusses implementations.

Abstract

Recent advancements in Natural Language Processing (NLP) use deep learning algorithms to improve performance. Google Translate shifts to neural machine translation, Baidu speech genetarion uses neural nets and question answering too. These neural networks share common architectures. They exploit recurrent computation to traverse the input and output. This talk will motivate the recurrent neural networks and discuss architectures. In the second half we discuss extensions such as attention mechanisms. Key words: RNN, seq2seq, attention, word vectors, data/model parallelism, low precision inference, TPU (explained in this order)

Subscribe to Receive PyData Updates

Subscribe

Tickets

Get Now