Memory Augmented
Neural Networks
Layton Hayes
Memory in Neural Networks: RNNs
(Recurrent Neural Networks)
Problems with RNNs
All because the memory is built into the network.
Memory outside Neural Networks: MANNs
(Memory Augmented Neural Network)
NTM (Neural Turing Machine):
NTMs continued: how they work
Addressing Mechanism
NTM: reading and writing basic equations
Mt -> the N x M memory matrix,
wt -> vector of weights, length N,
rt -> read vector
et -> erase vector, length M, range (0,1)
at -> add vector, length M
Reading:
Writing:
erase:
add:
NTM vs LSTM: Copy problem
Input:�Sequence of length L, �then nothing for L steps.
Output:�Nothing for L steps,�then repeat input sequence.
NTM vs LSTM: Copy problem generalization
Trained on copy problem for sequences of length L
results for sequences of length L, L*2, L*4, etc.
NTM:
LSTM:
NTM vs LSTM: repeat copy problem
Input:�Sequence of length L, �number of repeats (X),�then nothing for L * X steps.
Output:�Nothing for L+1 steps,�then repeat input sequence�X times.
Exciting applications: one-shot learning
Solution: don’t just learn; learn how to learn first.
One-shot Learning with Memory-Augmented Neural Networks
Input: Image, class of previous image
Output: class of current image
classes used, labels for each class, and specific samples are all shuffled between episodes
NTM++: Differentiable Neural Computer (DNC)
DNC: more detail
DNC: task demonstrations
bAbI
Issues with MANNs
Source Papers
Neural Turing Machines (Dec 2014)
https://arxiv.org/abs/1410.5401
One-shot Learning with Memory-Augmented Neural Networks (May 2016)
https://arxiv.org/abs/1605.06065
Hybrid computing using a neural network with dynamic external memory (Oct 2016)
https://www.nature.com/articles/nature20101