This document discusses various techniques for data augmentation of sound data. It begins by explaining that data augmentation increases data volume by deforming inputs in ways that could occur in the real world, such as flipping or pitch/time shifting, as long as it does not affect features needed for classification. Then, it focuses on techniques used for tasks like speech recognition and environmental sound classification, including traditional audio augmentations like adding noise or reverberation. Recent approaches discussed are SpecAugment, which drops sections of spectrograms, and Mixup/BC-learning, which create virtual samples between classes. It also mentions generating new data using cVAE or ACGAN models.