Pular para o conteúdo principal

Autoencoders variacionais (VAEs)

Definição

VAEs learn a latent space by training an encoder-decoder with a variational (reparameterized) objective. They support generation and smooth interpolation in latent space.

Eles se diferenciam de GANs (adversariais) e difusão (eliminação de ruído): o espaço latente é regularizado (KL to a prior) so it is smooth and interpretable. Generation can be blurrier than GANs/diffusion, but VAEs are useful for representation learning, anomaly detection, and when a low-D latent is desired.

Como funciona

Entrada é passada a um codificador que produz parâmetros de uma distribuição latente (por ex. média e log-variância par Gaussian). A z vector is sampled (reparameterization trick: z = mean + std * epsilon) and fed to the decoder, which reconstructs the input. Loss = reconstruction loss (por ex. MSE or cross-entropy) + KL divergence from the latent to a prior (por ex. standard normal). The KL term regularizes the latent space; the reconstruction term keeps it informative. At generation time, sample z from the prior and run the decoder.

Casos de uso

VAEs suit tasks that need a continuous latent space: smooth generation, anomaly detection, or learned representations.

  • Generative modeling with smooth latent interpolation
  • Anomaly detection via reconstruction error
  • Learned representations for downstream tasks

Documentação externa

Veja também