GANs. Machine Learning: Jordan Boyd-Graber University of Maryland SLIDES ADAPTED FROM GRAHAM NEUBIG

Similar documents
Language Models. Data Science: Jordan Boyd-Graber University of Maryland SLIDES ADAPTED FROM PHILIP KOEHN

Generative adversarial networks

Gradient descent GAN optimization is locally stable

Introduction to Machine Learning

Introduction to Machine Learning

Introduction to Machine Learning

Discrete Probability Distributions

Online Learning. Jordan Boyd-Graber. University of Colorado Boulder LECTURE 21. Slides adapted from Mohri

Logistic Regression. Introduction to Data Science Algorithms Jordan Boyd-Graber and Michael Paul SLIDES ADAPTED FROM HINRICH SCHÜTZE

Energy-Based Generative Adversarial Network

Generative Adversarial Networks. Presented by Yi Zhang

Introduction to Machine Learning

Machine Learning. Machine Learning: Jordan Boyd-Graber University of Maryland REINFORCEMENT LEARNING. Slides adapted from Tom Mitchell and Peter Abeel

Classification: The PAC Learning Framework

Nishant Gurnani. GAN Reading Group. April 14th, / 107

Classification: Logistic Regression from Data

Annotation and Feature Engineering

Chapter 20. Deep Generative Models

Generative Adversarial Networks

Machine Learning Basics Lecture 2: Linear Classification. Princeton University COS 495 Instructor: Yingyu Liang

Classification. Jordan Boyd-Graber University of Maryland WEIGHTED MAJORITY. Slides adapted from Mohri. Jordan Boyd-Graber UMD Classification 1 / 13

Reinforcement Learning for NLP

Inexact Search is Good Enough

Enforcing constraints for interpolation and extrapolation in Generative Adversarial Networks

Generative Adversarial Networks

Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks

Random Variables and Events

Introduction to Gaussian Process

Expectations and Entropy

Generative Adversarial Networks (GANs) Ian Goodfellow, OpenAI Research Scientist Presentation at Berkeley Artificial Intelligence Lab,

Machine Learning Linear Classification. Prof. Matteo Matteucci

Lecture 14: Deep Generative Learning

What s so Hard about Natural Language Understanding?

Towards a Data-driven Approach to Exploring Galaxy Evolution via Generative Adversarial Networks

COMP90051 Statistical Machine Learning

Robustness in GANs and in Black-box Optimization

Ian Goodfellow, Staff Research Scientist, Google Brain. Seminar at CERN Geneva,

Clustering. Introduction to Data Science. Jordan Boyd-Graber and Michael Paul SLIDES ADAPTED FROM LAUREN HANNAH

Decision Trees. Data Science: Jordan Boyd-Graber University of Maryland MARCH 11, Data Science: Jordan Boyd-Graber UMD Decision Trees 1 / 1

Support Vector Machines

Logistic Regression & Neural Networks

Logistic Regression. INFO-2301: Quantitative Reasoning 2 Michael Paul and Jordan Boyd-Graber SLIDES ADAPTED FROM HINRICH SCHÜTZE

GAUSSIAN PROCESS REGRESSION

The Success of Deep Generative Models

GENERATIVE ADVERSARIAL LEARNING

Deep Generative Models for Graph Generation. Jian Tang HEC Montreal CIFAR AI Chair, Mila

A Unified View of Deep Generative Models

Gaussians Linear Regression Bias-Variance Tradeoff

Generative Adversarial Networks, and Applications

Singing Voice Separation using Generative Adversarial Networks

MMD GAN 1 Fisher GAN 2

Do you like to be successful? Able to see the big picture

Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 5

Machine Learning Summer 2018 Exercise Sheet 4

Deep Learning Basics Lecture 8: Autoencoder & DBM. Princeton University COS 495 Instructor: Yingyu Liang

Deep Generative Models. (Unsupervised Learning)

Density Estimation. Seungjin Choi

Second Midterm Exam Economics 410 Thurs., April 2, 2009

CSCI-567: Machine Learning (Spring 2019)

Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 9

Name: Student number:

Sequence labeling. Taking collective a set of interrelated instances x 1,, x T and jointly labeling them

Wasserstein GAN. Juho Lee. Jan 23, 2017

Dimension Reduction (PCA, ICA, CCA, FLD,

Variational Autoencoder

Logistic Regression. Jia-Bin Huang. Virginia Tech Spring 2019 ECE-5424G / CS-5824

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Variational Autoencoders (VAEs)

Text2Action: Generative Adversarial Synthesis from Language to Action

Machine Learning

GANS for Sequences of Discrete Elements with the Gumbel-softmax Distribution

ECE 5984: Introduction to Machine Learning

Linear classifiers: Logistic regression

arxiv: v1 [cs.lg] 6 Nov 2016

Need for Sampling in Machine Learning. Sargur Srihari

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

text classification 3: neural networks

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation

Stochastic Backpropagation, Variational Inference, and Semi-Supervised Learning

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

Slide credit from Hung-Yi Lee & Richard Socher

Lecture 16 Deep Neural Generative Models

Energy Based Models. Stefano Ermon, Aditya Grover. Stanford University. Lecture 13

CENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 14. Sinan Kalkan

Probabilistic clustering

An Introduction to Statistical and Probabilistic Linear Models

Cycle-Consistent Adversarial Learning as Approximate Bayesian Inference

LDA with Amortized Inference

Topics in Natural Language Processing

Discriminative part-based models. Many slides based on P. Felzenszwalb

Hidden Markov Models

Graphical Models for Collaborative Filtering

ECE521 Lecture7. Logistic Regression

Machine Learning

arxiv: v3 [cs.lg] 2 Nov 2018

Log-linear models (part 1)

Generative Adversarial Networks (GANs) for Discrete Data. Lantao Yu Shanghai Jiao Tong University

Posterior Regularization

arxiv: v1 [stat.ml] 19 Jan 2018

Feedforward Neural Networks

Transcription:

GANs Machine Learning: Jordan Boyd-Graber University of Maryland SLIDES ADAPTED FROM GRAHAM NEUBIG Machine Learning: Jordan Boyd-Graber UMD GANs 1 / 7

Problems with Generation Generative Models Ain t Perfect Over-emphasis of common outputs, fuzziness Real MLE Adversarial Note: this is probably a good idea if you are doing (Lotter et al. maximum 2015) likelihood! Image Credit: Lotter et al. 2015 Fitting conventional prob models focuses on common input Can be fuzzy Still better for smaller ammounts of data or if true objective is ML Machine Learning: Jordan Boyd-Graber UMD GANs 2 / 7

Adversarial Training It s time for some game theory Machine Learning: Jordan Boyd-Graber UMD GANs 3 / 7

Adversarial Training It s time for some game theory Create discriminator that criticizes generated output Is this example real or not Generator is trained to fool discriminator to say it s real Machine Learning: Jordan Boyd-Graber UMD GANs 3 / 7

Adversarial Training It s time for some game theory Create discriminator that criticizes generated output Is this example real or not Generator is trained to fool discriminator to say it s real Contrast with encoder / decoder: Machine Learning: Jordan Boyd-Graber UMD GANs 3 / 7

Adversarial Training It s time for some game theory Create discriminator that criticizes generated output Is this example real or not Generator is trained to fool discriminator to say it s real Contrast with encoder / decoder: no fixed representation Machine Learning: Jordan Boyd-Graber UMD GANs 3 / 7

Training GAN Training Method sample latent vars. z sample minibatch convert w/ generator xreal xfake predict w/ discriminator discriminator loss (higher if fail predictions) y generator loss (higher if make predictions) Machine Learning: Jordan Boyd-Graber UMD GANs 4 / 7

Training Equations Discriminator l D (θ D,θ G ) = x P data [logd(x)] z [log(1 D(G(z)))] Real data should get high score Fake data should get low score Machine Learning: Jordan Boyd-Graber UMD GANs 5 / 7

Training Equations Discriminator l D (θ D,θ G ) = x P data [logd(x)] z [log(1 D(G(z)))] Real data should get high score Fake data should get low score Machine Learning: Jordan Boyd-Graber UMD GANs 5 / 7

Training Equations Discriminator l D (θ D,θ G ) = x P data [logd(x)] z [log(1 D(G(z)))] Real data should get high score Fake data should get low score Machine Learning: Jordan Boyd-Graber UMD GANs 5 / 7

Training Equations Discriminator Generator l D (θ D,θ G ) = x P data [logd(x)] z [log(1 D(G(z)))] Real data should get high score Fake data should get low score l G (θ D,θ G ) = l D (θ D,θ G ) If discriminator is very accurate, sometimes better to focus on non-saturating loss Focus on where you can confuse discriminator z [ logd(g(z))] (1) Machine Learning: Jordan Boyd-Graber UMD GANs 5 / 7

Problems with Training GANs are great, but training very hard Mode Collapse: generator maps all z to single x Over-confident discriminator Machine Learning: Jordan Boyd-Graber UMD GANs 6 / 7

Problems with Training GANs are great, but training very hard Mode Collapse: generator maps all z to single x (other examples as side information) Over-confident discriminator Machine Learning: Jordan Boyd-Graber UMD GANs 6 / 7

Problems with Training GANs are great, but training very hard Mode Collapse: generator maps all z to single x (other examples as side information) Over-confident discriminator (smoothing) Machine Learning: Jordan Boyd-Graber UMD GANs 6 / 7

Problem! Can t Backprop Problems with Discrete Data through Sampling sample latent vars. z sample minibatch convert w/ generator xreal predict w/ discriminator y xfake Discrete! Can t backprop Machine Learning: Jordan Boyd-Graber UMD GANs 7 / 7