Naoki Yokoyama
About MeResearchTutorialsSide Projects CV

An Unsupervised Generative Model

  • One of the most simple neural nets that serves as a generator, as opposed to a discriminator.
  • Encodes its input to a smaller dimension, then decodes that compressed representation to try to reconstruct the input.
  • Since the goal is to simply reconstruct the input, the training can be unsupervised, as the net already has the input to compare against the output.
  • Why reconstruct the input?

  • After training, the net will presumably have learned how to extract the important features of the input, and use them to recreate a convincing reconstruction.
  • Thus, if we pass a noisy input, the autoencoder should be able to accurately recreate the original input.
  • To make the autencoder more robust against certain types of noise, the noise could be added to the training samples before being passed through the net, but the output would still be compared against the original, noise-free input.
  • Noise Removal and Beyond

    Face completion using an AE (within a GAN) ["Generative Adversarial Denoising Autoencoder for Face Completion"]

    Semantic inpainting on masked images ["Context Encoders: Feature Learning by Inpainting" Pathak et al. CVPR 2016]

    Two autoencoders used for swapping faces between two people (actress' face replaced by Nicholas Cage's) (aka DeepFakes)