tracking the world state with recurrent entity networks
play

Tracking the World State with Recurrent Entity Networks Mikael - PDF document

11/7/2017 Tracking the World State with Recurrent Entity Networks Tracking the World State with Recurrent Entity Networks Mikael Henaff, Jason Weston, Arthur Szlam, Antoine Bordes, Yann LeCun Task At each timestep, get information (in the form


  1. 11/7/2017 Tracking the World State with Recurrent Entity Networks Tracking the World State with Recurrent Entity Networks Mikael Henaff, Jason Weston, Arthur Szlam, Antoine Bordes, Yann LeCun Task At each timestep, get information (in the form of a sentence) about the state of the world. Then answer a question. When we get new information, we should update our representation of the world state. The world state can be decomposed into the state of each entity in the world, so we only need to update one entity. Architecture The memory model: s , ⋯ s input: a sequence of vectors 1 T h , ⋯ h output: a set of entity representations 1 k The world is a collection of entities. Information about each entity is stored in a single cell. Each cell comes with a key and a memory slot. ~ h , w , s and g depends on h standard gating mechanism: https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 1/10

  2. 11/7/2017 Tracking the World State with Recurrent Entity Networks multiple-cells at multiple timesteps: https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 2/10

  3. 11/7/2017 Tracking the World State with Recurrent Entity Networks Input Encoder: input: a sequence of sentences. output: an encoding of each sentence as a fixed sized vector e are pretrained embeddings i Output Module: input: a query vector q and the outputs of the memory model output: arbitrary vector (log probabilities over words) https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 3/10

  4. 11/7/2017 Tracking the World State with Recurrent Entity Networks Look at only one entity and drop the query: y = Rϕ ( Hh ) = Rϕ j j y = R ϕ i i j Key vectors the model should identify entities by keys, which are trainable Key tying: Use parser to identify entities. One memory cell for each entity. Freeze key vector to be word embedding of an entity. Related work LSTM/GRU RENN https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 4/10

  5. 11/7/2017 Tracking the World State with Recurrent Entity Networks scalar memory cell with full interaction separate memory cells just sigmoid layer of input and hidden state content-based term between input and hidden state LSTM: Forget gate layer: Input gate layer & tanh(hyperbolic tangent) layer: RENN Memory Network store the entire input sequence in �dynamic long-term� a fixed number of blocks �a window of words� of hidden states memory as memories sequentially update a controller's hidden state via update each block with an independent gated RNN a softmax gating over the memories Gated graph network RENN inter-network communication with edges parallel/independent recurrent models Compared to RENN, CommNN, Interaction Network, Neural Physics Engine use parallel recurrent models without gating mechanism. Experiments Synthetic world model task Task details: Two agents randomly placed in a 10x10 grid Answer the locations of the agents based on up to T-2 supporting facts Details: 5 memory slots 20D per cell https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 5/10

  6. 11/7/2017 Tracking the World State with Recurrent Entity Networks bAbI Details: 20 memory cells 100D embedding U = V = 0, W = I , ϕ = id entity matrix https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 6/10

  7. 11/7/2017 Tracking the World State with Recurrent Entity Networks https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 7/10

  8. 11/7/2017 Tracking the World State with Recurrent Entity Networks Interpreting representations Recall that y = R ϕ i i j Find closest R for each entity ϕ : i j https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 8/10

  9. 11/7/2017 Tracking the World State with Recurrent Entity Networks CBT Input: �. 20 sentences �. 21st sentence with missing word �. list of candidate words Details: Tied keys to candidate words Dropout U = V = 0, W = I , ϕ = id No normalization https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 9/10

  10. 11/7/2017 Tracking the World State with Recurrent Entity Networks https://paper.dropbox.com/doc/print/RqoUdaN3Xe14IOEXWr1ze?print=true 10/10

Recommend


More recommend