Automated Tuning of Ad Auctions

Similar documents
COUNTERFACTUAL REASONING AND LEARNING SYSTEMS LÉON BOTTOU

Counterfactual Evaluation and Learning

Multiple-step Time Series Forecasting with Sparse Gaussian Processes

Counterfactual Model for Learning Systems

Counterfactual Reasoning and Learning Systems

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN SOLUTIONS

A Structural Model of Sponsored Search Advertising Auctions

Outline. Offline Evaluation of Online Metrics Counterfactual Estimation Advanced Estimators. Case Studies & Demo Summary

Gaussian Processes (10/16/13)

Deep Counterfactual Networks with Propensity-Dropout

MTTTS16 Learning from Multiple Sources

Introduction to Gaussian Processes

Integrated Electricity Demand and Price Forecasting

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2

Learning Gaussian Process Models from Uncertain Data

Large-scale Ordinal Collaborative Filtering

Self Adaptive Particle Filter

Standard Error of Technical Cost Incorporating Parameter Uncertainty

Probabilistic Graphical Models

Uncertainty in energy system models

Payment Rules for Combinatorial Auctions via Structural Support Vector Machines

Offline Evaluation of Ranking Policies with Click Models

Probabilistic Graphical Models Lecture 20: Gaussian Processes

Confidence Estimation Methods for Neural Networks: A Practical Comparison

Review: Probabilistic Matrix Factorization. Probabilistic Matrix Factorization (PMF)

Lecture 5: GPs and Streaming regression

Notes for Math 324, Part 19

Learning to Bid Without Knowing your Value. Chara Podimata Harvard University June 4, 2018

Prediction of Data with help of the Gaussian Process Method

Model Selection for Gaussian Processes

Adaptive Sampling Under Low Noise Conditions 1

Comparison of Modern Stochastic Optimization Algorithms

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression

Local and global sparse Gaussian process approximations

Lecture Notes on the Gaussian Distribution

Word-length Optimization and Error Analysis of a Multivariate Gaussian Random Number Generator

Software Tools: Congestion Management

Statistical Techniques in Robotics (16-831, F12) Lecture#21 (Monday November 12) Gaussian Processes

Learning to Learn and Collaborative Filtering

Cognitive Cyber-Physical System

Quantifying mismatch in Bayesian optimization

Machine Learning (CS 567) Lecture 2

A COLLABORATIVE 20 QUESTIONS MODEL FOR TARGET SEARCH WITH HUMAN-MACHINE INTERACTION

On the strategic use of quality scores in keyword auctions Full extraction of advertisers surplus

Towards a Bayesian model for Cyber Security

About Nnergix +2, More than 2,5 GW forecasted. Forecasting in 5 countries. 4 predictive technologies. More than power facilities

Probabilistic Graphical Models

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling

A SELF-TUNING KALMAN FILTER FOR AUTONOMOUS SPACECRAFT NAVIGATION

26 : Spectral GMs. Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G.

Generative Learning algorithms

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

Gaussian Process Regression Networks

Optimal Secondary Spectrum Auctions for Markets with Communication Constraints

Algorithms other than SGD. CS6787 Lecture 10 Fall 2017

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian with mean ( µ ) and standard deviation ( σ)

Decision making and problem solving Lecture 1. Review of basic probability Monte Carlo simulation

Reliability Monitoring Using Log Gaussian Process Regression

20: Gaussian Processes

An Adaptive Algorithm for Selecting Profitable Keywords for Search-Based Advertising Services

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

Statistical Techniques in Robotics (16-831, F12) Lecture#20 (Monday November 12) Gaussian Processes

Basics of reinforcement learning

WEATHER DEPENENT ELECTRICITY MARKET FORECASTING WITH NEURAL NETWORKS, WAVELET AND DATA MINING TECHNIQUES. Z.Y. Dong X. Li Z. Xu K. L.

Probabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm

Model-Based Reinforcement Learning with Continuous States and Actions

In the following panel I have added (in red) the proviso that X ~ p (as used on the left side of º):

Discrete Variables and Gradient Estimators

Machine Learning for Large-Scale Data Analysis and Decision Making A. Week #1

Glasgow eprints Service

Probability Distributions: Continuous

diluted treatment effect estimation for trigger analysis in online controlled experiments

A Bayesian Perspective on Residential Demand Response Using Smart Meter Data

Understanding Covariance Estimates in Expectation Propagation

Learning Non-stationary System Dynamics Online using Gaussian Processes

Multi-Kernel Gaussian Processes

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Essentials of expressing measurement uncertainty

Computational and Monte-Carlo Aspects of Systems for Monitoring Reliability Data. Emmanuel Yashchin IBM Research, Yorktown Heights, NY

Notifications and Accretions: Facility Role

Combine Monte Carlo with Exhaustive Search: Effective Variational Inference and Policy Gradient Reinforcement Learning

A BAYESIAN APPROACH FOR PREDICTING BUILDING COOLING AND HEATING CONSUMPTION

Generalized Linear Models in Collaborative Filtering

NetBox: A Probabilistic Method for Analyzing Market Basket Data

Online Supplementary Material. MetaLP: A Nonparametric Distributed Learning Framework for Small and Big Data

Lecture 4 January 23

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

A Process over all Stationary Covariance Kernels

CS 573: Algorithmic Game Theory Lecture date: April 11, 2008

A Structural Model of Sponsored Search Advertising Auctions

A Novel Click Model and Its Applications to Online Advertising

Answer Key for M. A. Economics Entrance Examination 2017 (Main version)

Real-Time Optimization (RTO)

Mixtures of Gaussians with Sparse Regression Matrices. Constantinos Boulis, Jeffrey Bilmes

Display Advertising Optimization by Quantum Annealing Processor

The Multivariate Gaussian Distribution [DRAFT]

Factor Modeling for Advertisement Targeting

RAO-BLACKWELLISED PARTICLE FILTERS: EXAMPLES OF APPLICATIONS

Transcription:

Automated Tuning of Ad Auctions Dilan Gorur, Debabrata Sengupta, Levi Boyles, Patrick Jordan, Eren Manavoglu, Elon Portugaly, Meng Wei, Yaojia Zhu Microsoft {dgorur, desengup, leboyles, pajordan, ermana, elonp, mewe, yaozhu}@microsoft.com Abstract We present an application of counterfactual reasoning for completely automated tuning of online ad auctions. We discuss practical considerations for applying the approach on a major search engine and present results showing parity between a reliable simulation engine and our approach that is orders of magnitude more efficient computationally. 1 Online Ad Auctions Selection, pricing, and allocation of ads on search result pages is determined via an auction that runs in real time whenever a query is submitted to a search engine. In each instance, relevant ads are selected to participate and allocation and pricing rules are subsequently executed to determine the placement and click-contingent price for each ad. Varian [3] popularized an abstraction of this auction known as Generalized Second Price, whereby ads are allocated in descending order of rank score and priced such that the payment per click is equal to the minimum bid required to maintain its current position according to the rank score function. Varian describes the rank score of ad i with bid b i as S i (b i ) = q i b i, where q i is some advertiser dependent quality score. In this paper, we consider a more generalized form of the rank score function, parameterized by auxiliary parameters {λ j } K 1, that influence pricing and allocation via advertiser, user and query dependent quality scores {q j } K 1, namely S i (b i λ, q). The auxiliary parameters control the trade-off between different key performance indicators (KPIs) such as revenue (RPM), click through rate (CTR), quick back rate (QBR), and ad coverage. The parameter values can be adjusted to achieve a particular operating point. Changes in the models that feed into the auction as well as changes in the marketplace make it necessary for the parameters to be readjusted, a process we refer to as tuning. Typically, tens of experiments need to be regularly tuned per market per device per month with many markets worldwide. The number of tunings required per month and the dynamic nature of marketplace requirements makes this a daunting task. This work builds upon [1] and focuses on the practical aspects of utilizing counterfactual reasoning in an end-to-end automated ad auction tuning pipeline for a major search engine. The idea is to use a distribution over the parameters at serve time, which enables asking what-if questions and predicting what the expected KPIs would be given alternative marketplace parameter settings. Here, we focus on the practical aspects of this approach and refer the reader to [1] for details on the basic technical details. Quantifying and Controlling Randomization Cost We make use of randomized auction parameters which are then used in conjunction with importance sampling in order to perform counterfactual estimation. For simplicity, we restrict our attention to the case with proposal N (µ, Σ). Choice of the proposal distribution p for which to sample the 30th Conference on Neural Information Processing Systems (NIPS 016), Barcelona, Spain.

Figure 1: Quantifying randomization cost using Monte Carlo simulations. The choice of the proposal distribution should be guided by the desired amount of exploration and the amount of deviation from the non-randomized experiment s KPIs that can be afforded. Randomization cost can be reduced using parameter correlations. auction parameters from affects the exploration power of the method. Proposals with larger variance will allow a larger range of counterfactuals to be considered. However, randomizing the parameters associated with a real-time auction has an impact on the overall KPIs as well, so there is a trade-off that must be considered. A larger variance will result in parameter samples further away from the current operating point. If the KPIs are not linear functions of the parameters, then the KPIs will be impacted in some way with the effect of the impact (typically) increasing with the size of the variance. The correlation structure of the parameters will impact the KPIs as well. Specifically, it is natural to correlate parameters that impact KPIs similarly so that more samples are exploring interesting regions of the counterfactual space. The corresponding joint covariance can have a different impact on the KPIs than that of an independent covariance with the same marginals as the joint. We utilized an existing auction replay and user simulation platform to obtain simulated counterfactuals in order to compute the correlation structure of the marketplace parameters, estimate the impact of randomization on the KPIs and finally decide on a covariance for online experiments. For computing the correlation structure, we simulated counterfactual KPIs for a large grid of marketplace parameters. We applied rather lose constraints on the counterfactual KPIs to eliminate points that lead to very poor operating points. We treated the parameter settings for the remaining points as possibly viable for a tuning scenario, and computed the correlation on these points. Next, we tested the impact of randomization using independent and correlated distributions with increasing variance. Let f be our KPI of interest and x our auction parameters. If the marketplace parameter distribution is p(x) and the distribution of the KPI given a parameter setting is p(f x), then we are interested in the quantity E p(f) [f] = E p(f,x) [f] = f p(x)p(f x)dxdf = dxp(x) fp(f x)df = E p(x) [E p(f x) [f x]]. (1) The simulation platform performs its estimations given user queries by training user click models and applying these to simulated auctions and clicks with computationally intensive simulation models, providing a reliable estimate of the KPI: ˆf(x) E p(f x) [f x]. Given this approximation, we can construct a Monte Carlo estimate of the KPI with E p(f) [f] E p(x) [ ˆf(x)] ˆf(x i ). () x i p(x) We estimated the correlation structure Σ of the parameters using simulation data as explained above. For Σ 1, a diagonal covariance matrix, and Σ, a full covariance matrix, such that Σ 1 and Σ have the same marginals, we evaluated () for p 1 = N (µ, c Σ 1 ) and p = N (µ, c Σ ) for a range of values for c (c = 0 corresponds to no randomization). Figure 1 compares the cost of randomization for different values of c, for the two matrices, Σ 1 (left) and Σ (right). We found that, as expected, larger values of c correspond to larger impact on the KPIs. Additionally, we see that jointly randomizing the auction parameters reduces the cost of randomization.

Table 1: Randomization Impact Comparison for Online Experiments(percent change) RPM QBR Space Impression Ind. γ = 1 0.38-0.15 0.48 0.51 Ind. γ = 0.5 0.1 0.08 0. 0.8 Joint γ = 1-0.15 0.1 0.09 0.11 In order to have a tighter grip on the randomization cost, we also introduce a randomization rate parameter that can be used to reduce the impact of randomization on the KPIs. To this end, we specify the flight distribution as p = (1 γ)δ µ + γn(µ, Σ), so that the impact of randomization can be reduced by reducing γ. In this way, we can reduce the impact of randomization without reducing the size of our effective counterfactual search space (as would be the case with reducing c), at the cost of requiring more time to gather the same amount of randomized data. Finally, we ran online experiments and compared their KPIs to an unrandomized baseline (with all means and centers the same). See Table 1. This experiment verifies that joint randomization indeed has lower impact than independent randomization. Some KPIs are closer to baseline even compared to independent randomization run with γ = 0.5. Thus, we conclude that utilizing correlation structure of the parameters leads to a more efficient exploration. 3 Estimation Uncertainty We use importance sampling to estimate counterfactual KPIs using a weighting of the KPIs observed using a particular sampling distribution in the online serving environment. Confidence intervals computed using importance sampling may be underestimated when the sample is not representative of the population. We use effective sample size as a diagnostic to tell if the importance sampling estimate is reliable. Noting that revenue or click based KPIs are generally larger variance than impression based KPIs, we compute both overall effective sample size and KPI-based effective sample size using ESS(f) = [ i w if(x i )] i w i f(x i), (3) and consider counterfactuals that have ESS lower than a threshold to be unreliable. 4 Increasing Computational Efficiency When both the proposal distribution and counterfactual distributions are Normally distributed with the same covariance, we can compute the ESS that would occur asymptotically for each counterfactual and pre-filter them before performing importance sampling. Consider samples x q = N(0, Σ x ) and a counterfactual distribution p = N(µ x, Σ x ), giving weights w(x) = p(x) q(x). Then we have log w(x) = x T Σ 1 x µ x µt x Σ 1 x As the above is a linear map on x and x N(0, Σ x ), we have log w(x) N( µt x Σ 1 x µ x µ x. (4), µ T x Σ 1 x µ x ). (5) Thus the weights are Log-Normal with mean parameter µ = µt x Σ 1 x µx and variance parameter σ = µ T x Σ 1 x µ x. As ) E[w] = exp (µ + σ = exp(0) = 1, E[w ] = exp ( µ + σ ) = exp ( µ T x Σ 1 x µ x ). Thus, asymptotically we expect the ESS on N samples to converge to ESS Ne µt x Σ 1 x µx. (6) 3

Figure : Comparing counterfactual estimates using IS to the simulations obtained from auction replay and user simulation platform. It is interesting to note that ESS depends on the shift in the mean, not on the dimension of the distribution. Thus, we can reliably evaluate the counterfactual KPIs for small changes to a large set of parameters as well as larger changes to a small set of parameters. We use this observation and asymptotic ESS values to guide choosing the search space and avoid computations on parameter settings that would not return reliable estimates. 5 Conclusion We have presented practical aspects of applying an efficient counterfactual estimation method utilizing importance sampling. It confers many advantages over other counterfactual estimation methods (simplicity of implementation, faster throughput, unbiasedness) at the cost of introducing some randomness into the real time auction. Comparison to the auction replay and user simulation platform in Figure shows that we produce similar estimates at a fraction of the computational cost. Using this approach, we can explore only a limited counterfactual space. However, the limited search space is not such a burden as tunings can be performed more frequently, even daily, in order to keep up with changing marketplace demands. Combined with a suitable KPI constrained optimization policy, we have run automated daily tunings. Early results show that we are able to successfully tune and achieve marketplace target KPIs. Such an automated workflow removes a human from the loop, improving the health and reliability of the tuning system, as it is less prone to human error. References [1] L. Bottou, J. Peters, J. Quiñonero Candela, D. X. Charles, D. M. Chickering, E. Portugaly, D. Ray, P. Simard, and E. Snelson. Counterfactual reasoning and learning systems: The example of computational advertising. Journal of Machine Learning Research, 14(Nov):307 360, 013. [] A. Swaminathan and T. Joachims. The self-normalized estimator for counterfactual learning. In C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems 8, pages 331 339. Curran Associates, Inc., 015. 4

[3] H. R. Varian. Position auctions. International Journal of Industrial Organization, 5(6):1163 1178, 006. 5