1. Unsupervised pretraining of deep neural networks (DNNs) on multiple datasets usually provides no benefit and sometimes harms performance compared to DNNs trained on individual datasets.
2. Representation learning aims to learn representations of input data that make a task easier by separating explanatory factors of variations. For example, mixture models can discover separate classes in data and distributed representations can divide the input space into uniquely identifiable regions.
3. Generative adversarial networks (GANs) can learn vector spaces that support semantic operations like representing a woman with glasses by combining vectors for concepts of gender and wearing glasses.