DeepMind's research presents a high-performance method for large-scale image recognition on ImageNet without the use of batch normalization, achieving state-of-the-art accuracy of 86.5%. The proposed adaptive gradient clipping method allows for efficient training with larger batch sizes, surpassing traditional models in terms of speed and performance. This study demonstrates that normalization-free networks can effectively be trained, showing better performance after finetuning on large datasets.