Combine Monte Carlo with Exhaustive Search: Effective Variational Inference and Policy Gradient Reinforcement Learning

Similar documents
Local Expectation Gradients for Doubly Stochastic. Variational Inference

Evaluating the Variance of

Stochastic Backpropagation, Variational Inference, and Semi-Supervised Learning

Bayesian Deep Learning

Variance Reduction in Black-box Variational Inference by Adaptive Importance Sampling

Using Gaussian Processes for Variance Reduction in Policy Gradient Algorithms *

Natural Gradients via the Variational Predictive Distribution

Variational Bayes on Monte Carlo Steroids

17 : Optimization and Monte Carlo Methods

The Generalized Reparameterization Gradient

VIME: Variational Information Maximizing Exploration

Bayesian Semi-supervised Learning with Deep Generative Models

Reinforcement Learning and NLP

Two Useful Bounds for Variational Inference

Overdispersed Black-Box Variational Inference

Probabilistic & Bayesian deep learning. Andreas Damianou

REBAR: Low-variance, unbiased gradient estimates for discrete latent variable models

Policy Gradient Methods. February 13, 2017

Deep Variational Inference. FLARE Reading Group Presentation Wesley Tansey 9/28/2016

Q-Learning in Continuous State Action Spaces

Learning Exploration/Exploitation Strategies for Single Trajectory Reinforcement Learning

Quasi-Monte Carlo Flows

Approximate Bayesian inference

REINTERPRETING IMPORTANCE-WEIGHTED AUTOENCODERS

Bayesian Networks BY: MOHAMAD ALSABBAGH

Sparse Approximations for Non-Conjugate Gaussian Process Regression

Policy Gradient Reinforcement Learning for Robotics

An Introduction to Expectation-Maximization

Variational Autoencoder

arxiv: v1 [stat.ml] 18 Oct 2016

arxiv: v2 [stat.ml] 15 Aug 2017

Lecture : Probabilistic Machine Learning

Introduction to Reinforcement Learning. CMPT 882 Mar. 18

A reinforcement learning scheme for a multi-agent card game with Monte Carlo state estimation

Doubly Stochastic Variational Bayes for non-conjugate Inference

Artificial Intelligence

Reparameterization Gradients through Acceptance-Rejection Sampling Algorithms

Understanding Covariance Estimates in Expectation Propagation

Reinforced Variational Inference

Lecture 8: Bayesian Estimation of Parameters in State Space Models

Bayesian Methods for Machine Learning

Lecture 16: Introduction to Neural Networks

Gradient Estimation Using Stochastic Computation Graphs

Black-box α-divergence Minimization

Auto-Encoding Variational Bayes

Bayesian Machine Learning

Evaluating the Variance of Likelihood-Ratio Gradient Estimators

A Process over all Stationary Covariance Kernels

Scalable Large-Scale Classification with Latent Variable Augmentation

Learning in Zero-Sum Team Markov Games using Factored Value Functions

Final Exam December 12, 2017

STA 4273H: Statistical Machine Learning

STA 4273H: Sta-s-cal Machine Learning

Approximate Inference Part 1 of 2

F denotes cumulative density. denotes probability density function; (.)

Reinforcement Learning II. George Konidaris

Approximate Inference Part 1 of 2

Bayesian Approach 2. CSC412 Probabilistic Learning & Reasoning

Variational Dropout via Empirical Bayes

Fast Inference and Learning for Modeling Documents with a Deep Boltzmann Machine

Reinforcement Learning II. George Konidaris

CLOSE-TO-CLEAN REGULARIZATION RELATES

Resampled Proposal Distributions for Variational Inference and Learning

Resampled Proposal Distributions for Variational Inference and Learning

Reparameterization Gradients through Acceptance-Rejection Sampling Algorithms

Auto-Encoding Variational Bayes. Stochastic Backpropagation and Approximate Inference in Deep Generative Models

PolicyBoost: Functional Policy Gradient with Ranking-Based Reward Objective

Nonparametric Bayesian Methods (Gaussian Processes)

Operator Variational Inference

1 Introduction 2. 4 Q-Learning The Q-value The Temporal Difference The whole Q-Learning process... 5

Multi-Attribute Bayesian Optimization under Utility Uncertainty

Variational Inference for Monte Carlo Objectives

Final Exam December 12, 2017

Parametric Inverse Simulation

Searching for the Principles of Reasoning and Intelligence

Reinforcement Learning as Variational Inference: Two Recent Approaches

Technical Details about the Expectation Maximization (EM) Algorithm

Denoising Criterion for Variational Auto-Encoding Framework

STA 4273H: Statistical Machine Learning

Expectation Propagation Algorithm

Study Notes on the Latent Dirichlet Allocation

Reinforcement Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Learning Gaussian Process Models from Uncertain Data

Reinforcement Learning for NLP

COMP3702/7702 Artificial Intelligence Lecture 11: Introduction to Machine Learning and Reinforcement Learning. Hanna Kurniawati

Reinforcement Learning

Planning by Probabilistic Inference

MCMC: Markov Chain Monte Carlo

Model-Based Reinforcement Learning with Continuous States and Actions

Bayesian Mixtures of Bernoulli Distributions

BACKPROPAGATION THROUGH THE VOID

Lecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions

CS Lecture 18. Expectation Maximization

LECTURE # - NEURAL COMPUTATION, Feb 04, Linear Regression. x 1 θ 1 output... θ M x M. Assumes a functional form

Stochastic Variational Inference for Gaussian Process Latent Variable Models using Back Constraints

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016

Making rating curves - the Bayesian approach

Variational Inference via Stochastic Backpropagation

bound on the likelihood through the use of a simpler variational approximating distribution. A lower bound is particularly useful since maximization o

Variational Sequential Monte Carlo

Transcription:

Combine Monte Carlo with Exhaustive Search: Effective Variational Inference and Policy Gradient Reinforcement Learning Michalis K. Titsias Department of Informatics Athens University of Economics and Business mtitsias@aueb.gr Abstract In this paper we discuss very preliminary wor on how we can reduce the variance in blac box variational inference based on a framewor that combines Monte Carlo with exhaustive search. We also discuss how Monte Carlo and exhaustive search can be combined to deal with infinite dimensional discrete spaces. Our method builds upon and extends a recently proposed algorithm that constructs stochastic gradients using local expectations. 1 Introduction Many problems in approximate variational inference and reinforcement learning involve the maximization of an expectation E qθ (x)[f(x), (1) where q θ (x) is a distribution that depends on parameters θ that we wish to tune. For most interesting applications in machine learning the optimization of the above cost function is very challenging since the gradient cannot be computed analytically. To deal with this, several approaches are based on stochastic optimization where stochastic gradients are used to carry out the optimization. Specifically, the two most popular approaches are the score function method (Williams, 1992; Glynn, 1990; Paisley et al., 2012; Ranganath et al., 2014; Mnih and Gregor, 2014) and the reparametrization method (Salimans and Knowles, 2013; Kingma and Welling, 2014; Rezende et al., 2014; Titsias and Lázaro-Gredilla, 2014). The reparametrization method is suitable for differential functions f(x) while the score function is completely general and it can be applied to non-smooth functions f(x) as well. Both these methods are heavily based on sampling from the variational distribution q θ (x). However, stochastic estimates purely based on Monte Carlo can be inefficient. This is because, while Monte Carlo has the great advantage that gives unbiased estimates it is computationally very intense and even in the simplest cases it requires infinite computational time to give exact answers. For instance, suppose x taes a finite set of K values. While the exact value of the gradient (based on exhaustive enumeration of all values) can be achieved in O(K) time, Monte Carlo still wastefully requires infinite computational time to provide the exact gradient. How can we reduce the computational ineffectiveness of Monte Carlo for (at least) discrete spaces? Here, we propose to do this by simply moving computational resources from Monte Carlo to exhaustive enumeration or search so that we construct stochastic estimates by combining Monte Carlo with exhaustive search. A first instance of such an algorithm has been recently proposed in (Titsias and Lázaro-Gredilla, 2015) and it is referred to as local expectation gradients or just local expectations. Here, we extend this approach by proposing i) a new version of the algorithm that can much more efficiently optimize highly correlated distributions q θ (x) where x is a high dimensional 29th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain.

discrete vector and ii) by showing how to deal with cases where components of x tae infinite values. In the appendix we also discuss applications to policy gradient optimization in reinforcement learning. 2 A new version of local expectations for correlated distributions Suppose that the n-dimensional latent vector x = (x 1,..., x n ) in the cost in (1) is such that each x i taes M i values. We consider a variational distribution over x that is represented as a directed graphical model having the following joint density n q θ (x) = q θ (x i pa i ), (2) i=1 where q θ (x i pa i ) is the conditional factor over x i given the set of the parents denoted by pa i and θ is the set of variational parameters. For simplicity we consider θ to be a global parameter that influences all factors, but in practice of course each factor could depend only on a subset of dimensions of θ. Next we will mae heavily use of a decomposition of q θ (x) in the form q θ (x) = q θ (x 1:i 1 )q θ (x i pa i )q θ (x i+1:n x i, x 1:i 1 ), where we intuitively thin of q θ (x 1:i 1 ) = i 1 j=1 q θ(x j pa j ) as the factor associated with the past, the single conditional q θ (x i pa i ) as the factor representing the present and the remaining term q θ (x i+1:n x i, x 1:i 1 ) = n j=i+1 q θ(x j pa j ) as the factor representing the future. To maximize the cost in (1) we tae gradients with respect to θ so that the exact gradient can be written as n θ E qθ (x)[f(x) = q θ (x 1:i 1 ) θ q θ (x i pa i ) q θ (x i+1:n x i, x 1:i 1 )f(x) i=1 x 1:i 1 xi x i+1:n To get an unbiased stochastic estimate we can firstly observe that the only problematic summation that is not already an expectation is the summation over x i. Therefore, our idea is to deal with the problematic summation over x i by performing exhaustive search, while for the remaining variables we can use Monte Carlo. We need two Monte Carlo operations that require sampling from the past factor q θ (x 1:i 1 ) and sampling also from the future factor q θ (x i+1:n x i, x 1:i 1 ). These two operations need to be treated slightly differently since the past factor is independent from x i, while the future factor does depend on x i. We can approximate the expectation under the past by drawing a sample x (s) 1:i 1 from q θ(x 1:i 1 ) yielding n M i θ E qθ (x)[f(x) θ q θ (x i = m pa (s) i ) q θ (x i+1:n x i = m, x (s) 1:i 1 )f(x(s) 1:i 1, x i:n) i=1 x i+1:n where we have explicitly written the sum over all possible values of x i. To get now an unbiased estimate of this we will need to draw M i samples from all possible future factors q θ (x i+1:n x i = m, x (s) 1:i 1 ), m = 1,..., M i which gives n M i θ E qθ (x)[f(x) θ q θ (x i = m pa (s) 1:i 1, x i = m, x (s,m) i+1:n ), (3) i=1 where x (s,m) i+1:n q θ(x i+1:n x i = m, x (s) 1:i 1 ). All these samples can be intuitively thought of as several possible imaginations of the future produced by exhaustively enumerating all values of x i at present time. Also the computations inside the sum n i=1 can be done in parallel. This is because we can a draw single sample x (s) from q θ (x) beforehand and then move along the path x (s) and compute all n terms involving the sums M i in parallel. The above algorithm is simpler and much more effective than the initial algorithm in (Titsias and Lázaro-Gredilla, 2015) which is based on the Marov blanet around x i, and it operates similarly to Gibbs sampling where a single future sample is used. Notice, however, that for fully factorised distributions where the past, present and future are independent the above procedure essentially becomes equivalent to local expectations in (Titsias and Lázaro-Gredilla, 2015). So it is only for the correlated distributions that the above stochastic gradients differ than the one proposed in (Titsias and Lázaro-Gredilla, 2015). In appendix we describe how the above algorithm can be used in reinforcement learning. 2

3 Dealing with infinite dimensional spaces For several applications, as those involving Bayesian non-parametric models or Poisson latent variable models, some random variables x i can tae countably infinite values. How can we extend the previous algorithm and the algorithms proposed in (Titsias and Lázaro-Gredilla, 2015) to deal with that? The solution is rather simple: we need to transfer computations from exhaustive search bac to Monte Carlo. Suppose that some random variable x i taes infinite values so in order to evaluate the stochastic gradient in (3) we need to perform the following infinite sum θq θ (x i = m pa (s) 1:i 1, x i = m, x (s,m) i+1:n ), which next in order to simplify notation we shall write as θ q θ (x i )f(x i ). We introduce a possibly adaptive integer T 1 and write the above as T q θ (x i )f(x i ) + q θ (x i )f(x i ). The first finite sum can be computed exactly through exhaustive enumeration. Thus, to get an overall unbiased estimate we need to get an unbiased estimate for the second term. For that we will use Monte Carlo. More precisely, to use Monte Carlo we need to apply the score function method so that the second term is written as q θ (x i )f(x i ) = q θ (x i ) log q θ (x i )f(x i ) = (1 Q(T )) = (1 Q(T )) q θ (x i ) 1 Q(T ) log q θ(x i )f(x i ) q vi (x i x i > T ) log q θ (x i )f(x i ) where Q(T ) = T q θ(x i ) is the cumulative distribution function and q θ (x i x i > T ) is the truncated variational distribution over the space x i > T. To get now an unbiased estimate we simply need to draw independent samples from q θ (x i x i > T ). So overall the stochastic gradient is T q θ (x i )f(x i ) + 1 Q(T ) S S s=1 log q θ (x (s) i ). (4) This gradient is unbiased for any value of T. However, in practice to efficiently reduce variance we will need to choose/adapt T so that the probability Q(T ) becomes large and the contribution of the second term is small. In practice, to implement this in a blac box manner we can adaptively choose T so that at each optimization iteration Q(T ) is above a certain threshold such as 0.95. 4 Discussion We have presented methods to combine Monte Carlo and exhaustive search in order to reduce variance in stochastic optimization of variational objectives and also to deal with infinite dimensional discrete spaces. The appendix describes a related algorithm for policy gradient reinforcement learning. References Glynn, P. W. (1990). Lielihood ratio gradient estimation for stochastic systems. Communications of the ACM, 33(10):75 84. Kingma, D. P. and Welling, M. (2014). Auto-encoding variational Bayes. In International Conference on Learning Representations. 3

Mnih, A. and Gregor, K. (2014). Neural variational inference and learning in belief networs. In International Conference on Machine Learning. Paisley, J. W., Blei, D. M., and Jordan, M. I. (2012). Variational Bayesian inference with stochastic search. In International Conference on Machine Learning. Ranganath, R., Gerrish, S., and Blei, D. M. (2014). Blac box variational inference. In Artificial Intelligence and Statistics. Rezende, D. J., Mohamed, S., and Wierstra, D. (2014). Stochastic bacpropagation and approximate inference in deep generative models. In International Conference on Machine Learning. Salimans, T. and Knowles, D. A. (2013). Fixed-form variational posterior approximation through stochastic linear regression. Bayesian Analysis, 8(4):837 882. Titsias, M. K. and Lázaro-Gredilla, M. (2014). Doubly stochastic variational Bayes for non-conjugate inference. In International Conference on Machine Learning. Titsias, M. K. and Lázaro-Gredilla, M. (2015). Local expectation gradients for blac box variational inference. In Advances in Neural Information Processing Systems. Williams, R. J. (1992). Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8(3 4):229 256. A Combine Monte Carlo and exhaustive search for policy optimization Consider a finite horizon Marov decision process (MDP) with joint density p(α 0:, s 1:h s 0) = π θ (α t s t)p(s s t, α t). (5) Let R = R(s t, α t, s ) denote the reward that the agent receives when starting at state s t, performing action α t and ending up at state s. The agent wishes to tune the policy parameters θ so that to maximize the expected total reward [ v θ (s 0) = E R where the expectation is taen under the distribution p(α 0:, s 1:h s 0) given by eq. (5). The gradient is explicitly written as [ θ v θ (s 0) = p(s 1:, α 0: 1 s 0) θ π θ (α s )p(s +1:h, α +1: s, α ) R α 0:,s 1:h where =0 = fθ (s 0) (6) =0 1 p(s 1:, α 0: 1 s 0) = π θ (α t s t)p(s s t, α t) p(s +1:h, α +1: s, α ) = p(s +1 s, α ) t=+1 π θ (α t s t)p(s s t, α t). The function f θ (s 0) can be thought as the gradient information collected by the agent at time, i.e. when the agent taes action α. The probability distribution p(s 1:, α 0: 1 s 0) describes the sequence of states and actions that precede action α, while p(s +1:h, α +1: s, α ) describes those generated after α is taen. Observe that while p(s 1:, α 0: 1 s 0) is independent from the current action α, p(s +1:h, α +1: s, α ) depends on it since clearly the future states and actions are influenced by the current action. The exact value of f θ (s 0) is computationally intractable and we are interested in expressing low variance unbiased estimates that can be realistically acquired through actual experience. The ey idea of our approach is 4

to explore all possible actions at time, and then combine this with Monte Carlo sampling. By taing advantage of the Marov structure of fθ (s 0) we re-arrange the summations as fθ (s 0) = [ p(s 1:, α 0: 1 s 0) θ π θ (α s ) p(s +1:h, α +1: s, α ) R s 1: α α s +1:h 0: 1 α +1: = s 1: α 0: 1 p(s 1:, α 0: 1 s 0) A(s ) θ π θ (α = m s ) p(s +1:h, α +1: s, α = m) s +1:h α +1: where ) denotes the number of possible actions when we are at state s(i). To get now an unbiased estimate we can sample from p(s 1:, α 0: 1 s 0) and from each conditional p(s +1:h, α +1: s, α = m). More precisely, by generating a single path (s (i) 1:, α(i) 0: 1 ) from the past factor p(s 1:, α 0: 1 s 0) we obtain ) θ π θ (α = m s (i) ) p(s +1:h, α +1: s s +1:h α +1: (i) [ R, α = m) [ R (i) Notice that the reward values R (i) are now indexed by i to emphasize their dependence on the drawn trajectory. However, the above quantity still remains intractable as the summation over all future trajectories, i.e. over the set (s +1:h, α +1: ), is not feasible. To deal with that we can use again Monte Carlo by sampling from each conditional MDP with distribution p(s +1:h, α +1: s (i), α = m) so that we start at state s (i), we tae action α = m and subsequently we follow the current policy. Thus, overall we generate ) total future trajectories, denoted by (s (i,m) +1:h, α(i,m) ) )A(s(i) ) +1:, and obtain the estimate θ π θ (α = m s (i) [ 1 ) R (i) + t= R (i,m) Here, is rather crucial to see that the sum of rewards is split into two different terms: 1 R(i) which is the sum of rewards obtained before the action α is taen, and t= R(i,m) which is the sum of rewards obtained by performing action α = m and then following the policy. Unlie the rewards in second term that depend on the current action (and therefore are indexed by both i and m), the first term is just a constant with respect to the action α. By taing advantage of that, the above unbiased estimate simplifies to ) θ π θ (α = m s (i) where we used that [ ) θ π θ (α = m s (i) ) 1 m s (i) ) = 0. [ ) t= R(i) R (i,m) = 1 R(i) (7) A(s (i) ) θ π θ (α = 5