This document discusses using gradient descent to learn update rules for optimization problems. It proposes learning the optimizer itself by treating it as a differentiable function with its own parameters. The goal is to learn update rules rather than using hand-designed rules. The method and results sections are missing details on how this is implemented and evaluated.