This document discusses backpropagation, an algorithm for supervised learning of artificial neural networks using gradient descent. It provides definitions and history of backpropagation, and explains how to use it with three main points:
1) It uses simple chain rules to calculate derivatives between weights in different layers to update weights.
2) Preparations include defining a cost function and the derivative of the sigmoid activation function commonly used.
3) The weight updates are dependent on derivatives from previous layers, and both forward and backward paths must be considered to calculate some derivatives between weights. Gradient descent is then applied to renew the weights.