Getting ready...

Now that we have the basic architecture of VAEs, the question arises how can they be trained, since the maximum likelihood of the training data and posterior density are intractable? The network is trained by maximizing the lower bound of the log data likelihood. Thus, the loss term consists of two components: generation loss, which is obtained from the decoder network through sampling, and the KL divergence term, also called the latent loss.

Generation loss ensures that the image generated by the decoder and the image used to train the network is the same, and latent loss ensures that the posterior distribution qᵩ(z|x) is close to the prior pϴ(z). Since the encoder uses Gaussian distribution for sampling, the latent loss measures how closely the latent variables match a unit Gaussian.

Once the VAE is trained, we can use only the decoder network to generate new images.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset