Questions tagged [autoencoder]

An autoencoder, autoassociator or Diabolo network is an artificial neural network used for learning efficient codings. As such, it is part of the dimensionality reduction algorithms.

The aim of an auto-encoder is to learn a compressed, distributed representation (encoding) for a set of data. This means it is being used for dimensionality reduction. Auto-encoders use two or more layers, starting from the input data (for instance in a face recognition task this would be pixels in the photograph):

  • A number of hidden layers (usually with a smaller number of neurons), which will form the encoder.
  • A number of hidden layer leading to an output layer (usually progressively bigger until the last one, where each neuron has the same meaning as in the input layer), which will form the decoder.

If linear neurons are used, then the optimal solution to an auto-encoder is strongly related to PCA.

When the hidden layers are larger than the input layer, an autoencoder can potentially learn the identity function and become useless; however, experimental results have shown that such autoencoders might still serve learn useful features in this case.

Auto-encoders can also be used to learn overcomplete feature representations of data.

The "coding" is also known as the embedded space or latent space in dimensionality reduction where the encoder will be used to project and the decoder to reconstruct.

1553 questions
0
votes
1 answer

tf.cond with tensors in one condition undefined

def get_z(epsilon): return tf.cond(flag,lambda: mean + sigma*epsilon,lambda: epsilon) In this, when I call the function with flag = True, I have my mean and sigma tensors defined and epsilon is the placeholder I give, and it works well. If I…
0
votes
2 answers

TensorFlow tfrecords: tostring() changes dimension of image

I have built a model to train a convolutional autoencoder in TensorFlow. I followed the instructions on Reading Data from the TF documentation to read in my own images of size 233 x 233 x 3. Here is my convert_to() function adapted from those…
0
votes
1 answer

(De-)Convutional lstm autoencoder - error jumps

I'm trying to build a convolutional lstm autoencoder (that also predicts future and past) with Tensorflow, and it works to a certain degree, but the error sometimes jumps back up, so essentially, it never converges. The model is as follows: The…
0
votes
1 answer

Feeding a hidden tensor in Tensorflow

I have an Some autoencoder. The model is not important now. Suppose that this model takes as input some image and output the reconstructed image. After training, I would like to see the effect of one tensor on the output. In addition, images are…
I. A
  • 2,252
  • 26
  • 65
0
votes
1 answer

Partially load tf.contrib.learn.Estimators from model_dir (load only decoder weights in auto-encoder setup)

I'm trying to work with Estimators rather than implementing training loops on my own. I'm playing with an auto-encoder on MNIST data. I have a training_model_fn function to build a training model that includes the inputs, model, loss, optimizers and…
Zach Moshe
  • 2,782
  • 4
  • 24
  • 40
0
votes
0 answers

tensorflow autoencoder implementation produces lots of noise in reconstruction

As mentioned in the title, I implemented a stacked autoencoder using tensorflow to compress image data into lower dimensions. However, the reconstructed image, after learning finished, contains static noise across the entire dataset. So I tried to…
0
votes
1 answer

Modifying Variational Autoencoder Architecture with SELU activation Function

In the variational auto encoder architecture we were using a mu and sigma fully connected layer. These are then followed by the latent variable layer, that samples from the Gaussian distribution, that is, from mu and sigma of the layer below. And in…
I. A
  • 2,252
  • 26
  • 65
0
votes
1 answer

How do I append label to image as input for Conditional VAE?

The rank is just not the same: [?, 28, 28, 1] and [?, 10]. How should I deal with this
Werma
  • 53
  • 1
  • 9
0
votes
1 answer

tensorflow autoencoder can not learn in training

In my code, I am trying to practice to use tr.train.batch function. In sess.run([optimizer]) line, it doesnt return anything and it is just frozen. Can you please find my mistake? tensors = tf.convert_to_tensor(x_train, dtype=tf.float32) tensors =…
user3104352
  • 1,100
  • 1
  • 16
  • 34
0
votes
1 answer

Classification with Keras Autoencoders

I'm trying to take a vanilla autoencoder using Keras (with a Tensorflow backend) and stop it when the loss value converges to a specific value. After the last epoch, I want to use a sigmoid function to perform classification. Would you know how to…
quil
  • 417
  • 1
  • 6
  • 16
0
votes
1 answer

python function does not work in Tensorflow: tf.contrib.layers.apply_regularization

my function is : def groupl1(x): return tf.reduce_sum(tf.sqrt(tf.to_float(x.get_shape()[1])) * tf.sqrt(tf.reduce_sum(x ** 2, axis=1))) and when i put it in my code: elif loss ==…
Joo Sohn
  • 25
  • 1
  • 6
0
votes
2 answers

tensorflow autoencoder is not learning in training

I am trying to generate synthetic data with tensorflow autoencoder that is very close to given original data. But, autoencoder is not learning in training phase. My cost function is not decreasing generally and synthetic data unrelated with original…
user3104352
  • 1,100
  • 1
  • 16
  • 34
0
votes
1 answer

Autoencoder parametric/non-parametric?

can you say that a autoencoder is parametric or non-parametric (in terms of pattern recognition) in general? I think both is possible because you can assume an underlying error distribution but you can also estimate the error with a different…
lymbot
  • 31
  • 6
0
votes
1 answer

Auto-encoder Mismatch Dimension Error

I am trying to reconstruct images using Conv Autoencoder, but I get an error related to dimensions, could you find a solution, thanks, Basically fist I want to test the model on reconstructing the same input data which are images, then if the model…
0
votes
1 answer

Fully convolutional autoencoder

I am implementing a convolutional autoencoder and I am having serious difficulty with finding the correct shapes for the convolution_transpose layers (in the decoder). So far my encoder looks like ('convolution', num_outputs=256, kernel_size=48,…
Qubix
  • 4,161
  • 7
  • 36
  • 73