About 50 results
Open links in new tab
  1. bayesian - What are variational autoencoders and to what learning …

    Jan 6, 2018 · 37 As per this and this answer, autoencoders seem to be a technique that uses neural networks for dimension reduction. I would like to additionally know what is a variational autoencoder …

  2. deep learning - When should I use a variational autoencoder as …

    Jan 22, 2018 · The standard autoencoder can be illustrated using the following graph: As stated in the previous answers it can be viewed as just a nonlinear extension of PCA. But compared to the …

  3. machine learning - Autoencoders as dimensionality reduction tools ...

    Jun 22, 2021 · As you mentioned in the question, Autoencoders serve as models which can reduce the dimensionality of their inputs. They are trained to "mimic" their inputs. The encoder produces a latent …

  4. What're the differences between PCA and autoencoder?

    Oct 15, 2014 · Both PCA and autoencoder can do demension reduction, so what are the difference between them? In what situation I should use one over another?

  5. neural networks - Why do we need autoencoders? - Cross Validated

    Recently, I have been studying autoencoders. If I understood correctly, an autoencoder is a neural network where the input layer is identical to the output layer. So, the neural network tries to pr...

  6. What is the difference between convolutional neural networks ...

    I can't tell you much about RBMs, but autoencoders and CNNs are two different kinds of things. An autoencoder is a neural network that is trained in an unsupervised fashion. The goal of an …

  7. Backpropagating regularization term in variational autoencoders

    Apr 1, 2025 · Backpropagating regularization term in variational autoencoders Ask Question Asked 11 months ago Modified 9 months ago

  8. PCA vs linear Autoencoder: features independence

    May 27, 2020 · The linear autoencoders in your question are not constrained to have an orthogonal basis, so we can't rely on this theorem when reasoning about the linear independence of the …

  9. neural networks - Deep Learning : Using dropout in Autoencoders ...

    Jun 10, 2018 · Autoencoders that include dropout are often called "denoising autoencoders" because they use dropout to randomly corrupt the input, with the goal of producing a network that is more …

  10. autoencoders - Should I be using batchnorm and/or dropout in a VAE …

    May 1, 2022 · I am trying to design some generative NN models on datasets of RGB images and was debating on whether I should be using dropout and/or batch norm. Here are my thoughts (I may be …