Dropout: Regularization for Neural Networks
Randomly dropping units during training to prevent overfitting
Keeping Neural Networks Simple by Minimizing the Description Length of the Weights
Hinton's MDL approach to neural network regularization through noisy weights
Recurrent Neural Network Regularization
How to apply dropout to LSTMs without disrupting memory dynamics