Generative adversarial networks (GANs) can be used to approximate the posterior distributions of Bayesian neural networks (BNNs). GANs are trained to generate samples from the posterior distribution learned by a BNN using stochastic gradient Langevin dynamics (SGLD). Specifically, a Wasserstein GAN with gradient penalty (WGAN-GP) is trained to match the posterior distribution by minimizing the Wasserstein distance between samples from the BNN's SGLD-approximated posterior and samples from the GAN's generator. This adversarial distillation technique allows parallel sampling from the BNN's posterior using only the GAN's parameters, providing computational and storage advantages over traditional Markov chain Monte Carlo methods for BNNs.