• Stars
    star
    150
  • Rank 247,323 (Top 5 %)
  • Language
    Python
  • Created over 8 years ago
  • Updated over 7 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

An implementation of an autoencoder with a learned similarity metric for reconstructing video frames

Learned Similarity Autoencoder for Modelling and Reconstructing Video Frames

Tesorflow implementation of Autoencoding beyond pixels using a learned similarity metric.

This was written for Tensorflow version 0.7 which is no longer supported. Unfortunately training a new model using Tensorflow 0.8 or later does not work as a model will learn at all. I believe this is an issue with the loss function of the discriminator and how it is backpropagated but I have been unable to debug this problem. If someone does fix it please submit a pull request.

A lot of the architecture is derived from this codebase DCGAN-Tensorflow.

This project is designed to read 256x144 png's and that are indexed in numerial order.

This project was implemented for my project on reconstructing videos with neural networks - read more

To train a model with a dataset:

$ python main.py --dataset DATASETNAME --is_train True 

You may want to adjust the amount of noise injected into the latent space:

$ python main.py --dataset DATASETNAME --is_train True --noise 0.5

This parameter controls the standard deviation of noise epsilon from mean of 0.

The output frames in sequence using an exisiting model:

$ python main.py --dataset DATASETNAME --is_run True 

In this version of the code 1000 minibatches must be processed before a model is saved, and therefore can be used to output frames sequentailly. This can be edited in line 245 of model.py.

Put the dataset directory in a directory called 'datasets' within the code project file.

To turn a video into frames and frame into video you can use ffmpeg