deep learning for language understanding software … fileconfidential + proprietary confidential +...

37
Confidential + Proprietary Confidential + Proprietary Deep Learning for Language Understanding (at Google Scale) Anjuli Kannan Software Engineer, Google Brain

Upload: doliem

Post on 04-Jun-2018

227 views

Category:

Documents


0 download

TRANSCRIPT

Confidential + ProprietaryConfidential + Proprietary

Deep Learning for Language Understanding (at Google Scale)Anjuli KannanSoftware Engineer, Google Brain

Confidential + Proprietary

Text is just a sequence of words

["hi", "team", "the", "server", "appears", "to", "be", "dropping", "about", "10%", …]

About me

● My team: Google Brain○ "Make machines intelligent, improve people's lives."○ Research + software + applications○ g.co/brain

● My work is at boundary of research and applications● Focus on natural language understanding

Neural network basics

Confidential + Proprietary

Neural network

Is a 4

Is a 5

...

...

Image: Wikipedia

Confidential + Proprietary

Neural network

Neuron

Is a 4

Is a 5

Confidential + Proprietary

Basic building block is the neuron

Greg Corrado

Gradient descent

w’ = w - α ∂wL(w)w

w’

Learning Rate

Slide: Vincent Vanhoucke

Recurrent neural networks

Confidential + Proprietary

Recurrent neural networks can model sequences

Recurrent neural networks can model sequences

How

Message

How are

Message

Recurrent neural networks can model sequences

How are you

Message

Recurrent neural networks can model sequences

How are you ?

Message

Recurrent neural networks can model sequences

Internal state is a fixed length encoding of the message

How are you ?

Message

Recurrent neural networks can model sequences

Sequence-to-sequence models

Suppose we want to generate email replies

SmartreplyIncoming email

Response email

Sequence-to-sequence model

Sutskever et al, NIPS 2014

Sequence-to-sequence model

encoder decoder

Sequence-to-sequence model

Ingests incoming message Generates reply message

Encoder ingests the incoming message

Internal state is a fixed length encoding of the message

How are you ?

Message

Decoder is initialized with final state of encoder

How are you ? __ How are you ?

Message

Decoder is initialized with final state of encoder

How are you ? __ How are you ?

Message

How are you ? __

I

Message

Response

Decoder predicts next word

How are you ? __ I

I am

Message

Response

Decoder predicts next word

How are you ? __ I am

I am great

Message

Response

Decoder predicts next word

How are you ? __ I am great

I am great !

Message

Response

Vinyals & Le, ICML DL 2015Kannan et al, KDD 2016

Decoder predicts next word

What the model can do

What the model can do

Summary

- Neural networks learn feature representations from raw data- Recurrent neural networks have statefulness which allows them to model

sequences of data such as text- The sequence-to-sequence model contains two recurrent neural networks: one

to encode an input sequence and one to generate an output sequence

Smartreply

Google Translate

Research: Speech recognition

Research: Electronic health records

What's next?

?

Resources

- All tensorflow tutorials: https://www.tensorflow.org/versions/master/tutorials/index.html

- Sequence-to-sequence tutorial (machine translation): https://www.tensorflow.org/versions/master/tutorials/seq2seq

- Chris Olah's blog: http://colah.github.io/

Thank you!