This document summarizes state-of-the-art algorithms for neural network quantization from Qualcomm AI Research. It introduces post-training quantization and quantization-aware training as two main approaches. Post-training quantization requires no retraining but can achieve 8-bit quantization with close to floating point accuracy. Quantization-aware training requires retraining but enables lower bit quantization with competitive results. The document provides tested pipelines for both approaches that achieve state-of-the-art performance on common models and tasks.