FNN-VAE for noisy time series forecasting

In the last part of this mini-series on forecasting with false nearest neighbors (FNN) loss, we replace the LSTM autoencoder from the previous post by a convolutional VAE, resulting in equivalent prediction performance but significantly lower training time. In addition, we find that FNN regularization is of great help when an underlying deterministic process is obscured by substantial noise.

Related Articles

Introduction to Autoencoders

Table of Contents Introduction to Autoencoders What Are Autoencoders? How Autoencoders Achieve High-Quality Reconstructions? Revisiting the Story Types of Autoencoder Vanilla Autoencoder Convolutional Autoencoder (CAE) Denoising Autoencoder Sparse Autoencoder Variational Autoencoder (VAE) Sequence-to-Sequence Autoencoder What Are the Applications of Autoencoders?…
The post Introduction to Autoencoders appeared first on PyImageSearch.

CycleGAN: Unpaired Image-to-Image Translation (Part 3)

Table of Contents CycleGAN: Unpaired Image-to-Image Translation (Part 3) Configuring Your Development Environment Need Help Configuring Your Development Environment? Project Structure Implementing CycleGAN Training Implementing Training Callback Implementing Data Pipeline and Model Training Perform Image-to-Image Translation Summary Citation Information CycleGAN:…
The post CycleGAN: Unpaired Image-to-Image Translation (Part 3) appeared first on PyImageSearch.