Keywords: Normalization Dropout Related Papers: Batch Normalization- Accelerating Deep Network Training by Reducing Internal Covariate Shift Layer Normalization Dropout- A Simple way to Prevent Neural Networks from Overfitting