This document summarizes the paper "Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift". It introduces batch normalization, which normalizes layer inputs to speed up training of neural networks. Batch normalization reduces internal covariate shift by normalizing layer inputs. It computes normalization statistics over each mini-batch and applies them to the inputs. This allows higher learning rates and acts as a regularizer. Experiments show batch normalization stabilizes and accelerates the training of neural networks on ImageNet classification.