Enforcing constraints for interpolation and extrapolation in Generative Adversarial Networks

Similar documents
Generative adversarial networks

Q-Learning in Continuous State Action Spaces

GENERATIVE ADVERSARIAL LEARNING

COMS 4771 Introduction to Machine Learning. Nakul Verma

Deep Feedforward Networks

Bayesian Deep Learning

Data Mining Part 5. Prediction

Generative Adversarial Networks

An Overview of Edward: A Probabilistic Programming System. Dustin Tran Columbia University

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Approximation Methods in Reinforcement Learning

Singing Voice Separation using Generative Adversarial Networks

Introduction to Reinforcement Learning. CMPT 882 Mar. 18

Probabilistic Graphical Models

Deep Generative Models for Graph Generation. Jian Tang HEC Montreal CIFAR AI Chair, Mila

Learning in State-Space Reinforcement Learning CIS 32

Reading Group on Deep Learning Session 1

Reinforcement Learning

Lecture 14: Deep Generative Learning

Deep Feedforward Networks

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Overview of gradient descent optimization algorithms. HYUNG IL KOO Based on

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

Generative Adversarial Networks, and Applications

Energy Based Models. Stefano Ermon, Aditya Grover. Stanford University. Lecture 13

Neural Networks and Deep Learning

Gradient descent GAN optimization is locally stable

Negative Momentum for Improved Game Dynamics

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Lecture 1: March 7, 2018

Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks

Generative Adversarial Networks for Real-time Stability of Inverter-based Systems

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

GANs. Machine Learning: Jordan Boyd-Graber University of Maryland SLIDES ADAPTED FROM GRAHAM NEUBIG

Today s s Lecture. Applicability of Neural Networks. Back-propagation. Review of Neural Networks. Lecture 20: Learning -4. Markov-Decision Processes

Chapter 20. Deep Generative Models

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Wasserstein GAN. Juho Lee. Jan 23, 2017

Quantum Artificial Intelligence and Machine Learning: The Path to Enterprise Deployments. Randall Correll. +1 (703) Palo Alto, CA

Robustness in GANs and in Black-box Optimization

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

Experiments on the Consciousness Prior

Comparison of Modern Stochastic Optimization Algorithms

Normalization Techniques in Training of Deep Neural Networks

Learning Particle Physics by Example:

Machine Learning. Boris

Summary of A Few Recent Papers about Discrete Generative models

Deep Learning Lab Course 2017 (Deep Learning Practical)

Nishant Gurnani. GAN Reading Group. April 14th, / 107

Summary and discussion of: Dropout Training as Adaptive Regularization

Learning Tetris. 1 Tetris. February 3, 2009

Lecture 16 Deep Neural Generative Models

Machine Learning and Bayesian Inference. Unsupervised learning. Can we find regularity in data without the aid of labels?

Self Supervised Boosting

Discrete Latent Variable Models

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

Learning Gaussian Process Models from Uncertain Data

The Deep Ritz method: A deep learning-based numerical algorithm for solving variational problems

Deep Generative Models. (Unsupervised Learning)

Effect of number of hidden neurons on learning in large-scale layered neural networks

Learning Control for Air Hockey Striking using Deep Reinforcement Learning

Logistic Regression & Neural Networks

Text2Action: Generative Adversarial Synthesis from Language to Action

Unsupervised Learning

Enhancing a Model-Free Adaptive Controller through Evolutionary Computation

Reinforcement Learning via Policy Optimization

Introduction of Reinforcement Learning

Reinforcement Learning

GANs, GANs everywhere

Serious limitations of (single-layer) perceptrons: Cannot learn non-linearly separable tasks. Cannot approximate (learn) non-linear functions

A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement

Training Generative Adversarial Networks Via Turing Test

Incremental Policy Learning: An Equilibrium Selection Algorithm for Reinforcement Learning Agents with Common Interests

Neural Networks. Yan Shao Department of Linguistics and Philology, Uppsala University 7 December 2016

Reinforcement Learning

CS Machine Learning Qualifying Exam

Machine Learning with Quantum-Inspired Tensor Networks

CSC321 Lecture 8: Optimization

Neural networks and optimization

Recap from previous lecture

CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning

Approximate Dynamic Programming

Recurrent Neural Network Training with Preconditioned Stochastic Gradient Descent

Lecture 23: Reinforcement Learning

Deep generative models for fast shower simulation in ATLAS

ECS171: Machine Learning

Grundlagen der Künstlichen Intelligenz

Machine Learning for Physicists Lecture 1

Policy Gradient Reinforcement Learning for Robotics

Towards a Data-driven Approach to Exploring Galaxy Evolution via Generative Adversarial Networks

Generative Adversarial Networks

Variational Autoencoders (VAEs)

Regularization and Optimization of Backpropagation

Generative Adversarial Networks. Presented by Yi Zhang

DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION. Alexandre Iline, Harri Valpola and Erkki Oja

The Success of Deep Generative Models

MMD GAN 1 Fisher GAN 2

attention mechanisms and generative models

Approximate inference in Energy-Based Models

Gaussian Processes (10/16/13)

Transcription:

Enforcing constraints for interpolation and extrapolation in Generative Adversarial Networks Panos Stinis (joint work with T. Hagge, A.M. Tartakovsky and E. Yeung) Pacific Northwest National Laboratory space space space Supported by Deep Learning for Scientific Discovery Investment" and PhILMs"

Statement of the problem Generative Adversarial Networks (GANs) are becoming popular machine learning choices for training generators. There is a concerted effort in the machine learning community: i) to expand the range of tasks in which learning can be applied and ii) to utilize methods from other disciplines to accelerate learning. Task: We want to enforce given constraints in the output of a GAN generator both for interpolation and extrapolation (prediction). Given a time series, we wish to train GAN generators to represent the flow map of a system. Remark: The cases of interpolation and extrapolation should be treated differently.

Definition and basic properties of GANs We are interested in training a generator to produce data from a given distribution (called the true data distribution). GANs formulate this task as a game between the generator and an associated discriminator (Goodfellow et al., 2014). The objective of the generator is to trick the discriminator into deciding that the generator-created samples actually come from the true data distribution. The best possible performance of the generator is to convince the discriminator half of the time that the samples it is generating come from the true distribution.

A two-player minimax game with value function V (D, G): min G max V (D, G) = E x p data (x)[log D(x)] + E D z pz(z)[log(1 D(G(z)))], where p data (x) is the true distribution and p z (z) is the input distribution of the generator. The generator G and the discriminator D are assumed to be neural networks with parameters θ g and θ d respectively. For a given generator G, the optimal discriminator DG (x) is given by D G (x) = p data (x) p data (x) + p g (x), where p g (x) is the generator s output distribution.

We can define C(G) = max D V (D, G) = V (D G, G). 1) The global minimum of C(G) is obtained if and only if p g = p data. 2) Since for this minimum we have DG (x) = 1 2, the value of the minimum is log 4. 3) For the generator, the minimum of log(1 D(G(z))) is log 2. Remark: GANs are notoriously difficult to train. In particular, the discriminator learns much faster than the generator and this leads to an instability of the training process. Remark: There are many variants of the basic GAN framework. Our construction can be applied to those variants too.

Enforcing constraints for interpolation We have a prior distribution p z (z) on the input of the generator and we would like to train the generator to produce samples from a function f (z) i.e., we want the data x = f (z). In the original implementation of a GAN, for each z we produce an x that satisfies the constraint x = f (z). We feed the discriminator both with the data x as well as the data x = G(z) that the generator produces. Remark: This implementation is straightforward but does not utilize the known bond between z and x. Idea: We can enforce the constraint by augmenting the discriminator input vector as (z, ɛ) where ɛ = x f (z) for the true sample. Similarly, for the generator-created sample, we augment the discriminator input vector as (z, ɛ ) where ɛ = x G(z).

Pros and cons The modification is easy to implement with modern software e.g. Tensorflow. We introduce a constraint for the data created by the generator. This constraint influences directly the parameters of the generator neural network (weights, biases in the case of convolutional neural nets) through back propagation. We respect the game-theoretic setup of GANs (under reasonable assumptions about the smoothness of the constraint). The modification can exacerbate the known instability issue of GANs.

Adding noise to promote stability To avoid the instability we can regulate the discriminator s ability to distinguish between generator-created and true data by adding noise to the constraint residual for the true data only. If the true data satisfy the constraint exactly, then we can decide how much noise to add based on Monte Carlo considerations. If the true data come from a numerical method (with given order of accuracy) or a physical experiment (with known precision), then we can use this information to decide the magnitude of the noise. The idea behind adding the noise is that the discriminator will become more tolerant by allowing generator-created samples for which the constraint residual is not exactly 0 but essentially within a narrow interval around it.

Numerical example - interpolation Game-theoretic error convergence 1.8 1.6 1.4 1.2 1.0 0.8 0.6 10 Discriminator error (constraint enforced) 1 Discriminator error (constraint not enforced) Generator error (constraint enforced) Generator error (constraint not enforced) 0.1 Relative error in logarithmic scale Constraint enforced (10 4 samples) Constraint not enforced (10 4 samples) 0 10,000 20,000 30,000 40,000 50,000 60,000 70,000 (a) Iterations 0 10,000 20,000 30,000 40,000 50,000 60,000 70,000 (b) Iterations Figure: Two linearly coupled oscillators. (a) Comparison of the evolution of the absolute value of the generator and discriminator game-theoretic error with and without enforcing a constraint (linear-linear plot) ; (b) Comparison of the evolution of the relative error RE m of the function learned with and without enforcing a constraint (linear-log plot). Remark: The game-theoretic optimum can be reached much faster than the actual relative error threshold (very different solution landscapes require adaptive learning rate)

Enforcing constraints for extrapolation 2.50 2.50 2.25 2.25 2.00 2.00 x_pos_extra 1.75 1.50 x_pos_extra 1.75 1.50 1.25 1.25 1.00 1.00 0.75 0.75 0.50 0 2 4 6 8 10 z 0.50 0 2 4 6 8 10 z (a) (b) Figure: Two linearly coupled oscillators. Comparison of target function x 1 (z) = R sin(z) (blue dots) and GAN prediction xpos extra (z) (red crosses). (a) 10 4 samples of noiseless data with the constraints enforced during training and a projection step; (b) 10 4 samples of noisy data with the constraints enforced during training and a projection step.

The need for error-correction Remark: The repeated application of the flow map neural network model (generative model) leads to error accumulation. Remark: The predicted trajectory veers off in untrained parts of phase space. As a result, the generator predictions fail. Idea: Introduce an error-correcting (restoring) force Train an interpolatory generator with noisy data Implementation: We center a cloud of points at each data point that is provided on the trajectory. Then, during training, at each step we force the GAN generator to map a point in this cloud to the correct (noiseless) point on the trajectory at the next step. Remark: This is akin to temporal renormalization.

Generator dynamics = Approximate dynamics + Restoring force Incorporate the restoring force implicitly by training with noisy data Incorporate the restoring force explicitly by training with noisy data and by learning the functional form of the restoring force Remark: The functional form of the restoring force can come from (temporal) model reduction considerations. Remark: It can even be represented by a separate neural network. Remark: The magnitude of the noise depends on the accuracy of the ground truth (order of accuracy, measurement errors). Remark: Expect scaling laws for parameters as a function of stepsize (incomplete similarity).

25 20 15 10 5 25 20 15 10 5 x1 0 x1 0 5 10 15 20 0 1 2 3 4 5 6 7 8 9 z (a) 5 10 15 20 0 1 2 3 4 5 6 7 8 9 z (b) Figure: Lorenz system. Comparison of ground truth x 1 (z) computed with the Euler scheme with stepsize δz = 10 4 (blue dots), the GAN prediction with stepsize z = 1.5 10 2 (red crosses) and the Euler scheme prediction with stepsize z = 1.5 10 2 (green triangles). (a) 2 10 4 samples of noisy data without enforced constraints during training; (b) 2 10 4 samples of noisy data with enforced constraints during training.

Current and future work We have obtained results for the construction both in the case of unsupervised learning (GANs) and supervised learning (add penalty terms to objective function). We will explore similar constructions for reinforcement learning. The role of noise is played there by the concept of stochastic exploration of the action policy space. Implementation in all 3 major models of learning allows comparison of the relative merits for each setting. This construction can be combined with online data acquisition to develop models that adapt to a changing environment. P. Stinis, T. Hagge, A.M. Tartakovsky and E. Yeung, Enforcing constraints for interpolation and extrapolation in Generative Adversarial Networks (2018) arxiv:1803.08182