This document summarizes knowledge distillation techniques for federated learning. It discusses how knowledge distillation has been used to enable model heterogeneity by exchanging model outputs or features instead of parameters. It also describes how distillation can be applied at the server-side to refine model aggregation or at the client-side to mitigate the effects of non-IID data distributions. The document structures the discussion according to whether distillation is used to allow for model heterogeneity or address data heterogeneity and provides examples of approaches within each category.