Why on earth did you do that?!
|
|
- Roderick Dickerson
- 5 years ago
- Views:
Transcription
1 Why on earth did you do that?! A formal model of obfuscation-based choice Caspar Chorus Delft University of Technology Challenge the future
2 Background: models of decision-making Based on the notion that choices are based on motivations (preferences, desires, decision-rules) - Theory of Reasoned Action / Planned Behavior (Psych.) - Optimal Decision-Making (OR, Dec. Sciences) - Rational Choice Theory (Economics, Discrete Choice) - Belief-Desire-Intention framework (Multi-Agent Systems, AI) - In other words, motivations echo through in choices e.g. Revealed preference axiom Cornerstone of micro-econom(etr)ics 2
3 Background: models of decision-making (II) Based on the notion that choices are based on motivations (preferences, desires, decision-rules) - Theory of Reasoned Action / Planned Behavior (Psych.) - Optimal Decision-Making (OR, Dec. Sciences) - Rational Choice Theory (Economics, Discrete Choice) - Belief-Desire-Intention framework (Multi-Agent Systems, AI) - In other words, motivations echo through in choices This talk: decision-maker wishes to suppress the echo In other words, wishes to hide his motivations from an onlooker 3
4 Example two futures for AI Dystopia AI outsmarts us Autonomous AI AI fights us AI out of control Opaque AI Unaligned AI Utopia AI outsmarts us Autonomous AI AI supports us AI in control Interpretable AI Aligned AI 4
5 Example (ctd): AI versus supervisor Dystopia AI fights us AI outsmarts us AI out of control Opaque AI Obfuscates rules Unaligned AI Utopia AI supports us AI outsmarts us AI in control Interpretable AI Tries to infer rules Aligned AI Actions are observeable, rules are latent 5
6 AI vs supervisor (illustration): algorithmic discrimination Artificial, autonomous agent Task: select CVs for job-interview (or: determine recidivism-probability) (or: set optimal health insurance premium) Rules: may include racist, sexist biases Human supervisor Aims to find out which rules agent uses Will punish use of biased rules Agent knows this and will try to obfuscate its rules 6
7 Some other ways to frame obfuscation Moral wiggle room & Human (Moral) decision making Privacy-protection - Human: facing a choice / moral dilemma - Onlooker: punishes with contempt (based on employed principle!) - Human: creates ambiguity about employed principle(s) Legal decision making - Suspect: making choice that (may) violate(s) law - Prosecutor: punishes with legal action (based on motivation!) - Suspect: creates reasonable doubt about motivations Military decision making - Autonomous weapon: infiltrates, fights - Adversary (human, AI): wins, if it finds out strategic objectives - Autonomous weapon: creates strategic ambiguity 7
8 Relevance: variety of research fields Supervisor Obfuscator Human Artificial Human Artificial Artificial Intelligence, (moral) Psychology, Law Expert Systems, Human-Computer Interaction Artificial Intelligence, Expert Systems, Multi-agent sytems Human-Computer Interaction 8
9 Relevance: also in Transport Supervisor Obfuscator Human Artificial Human Moral decisions by travelers and citizens (e.g. taboo trade-offs) Keeping your Automated Vehicle on the moral high Artificial Travel recommender system (e.g. personal travel assistant) AV-AV interaction on multi-lane highways, ground ( moral machine ) crossroads 9
10 Intermezzo: obfuscation vs deception Why not assume that agents try to mislead? (rather than assuming that they try to create a smoke screen) Because agent may not be bad, merely afraid of punishment And creating reasonable doubt is enough to avoid it Because deceit is costly when found out (compared to smoke) And obfuscating is much harder to proove than deceit Because agent does not know which rule is considered right one And as such has no target rule for deceit 1
11 Base Model 11
12 The model basic notation Set contains actions (alternatives, options) Set contains rules (motivations) by matrix contains scores describing how an action performs on a given rule. +,, : obliged (+), permitted (), prohibited ( ) Strong rule: +, Weak rule:, 12
13 Behavior of a Rule-follower Agent follows rule. Probability that she chooses given that she follows : Strong rule = 1 if is obliged under = if is prohibited under Weak rule = 1 if is permitted (but not obliged) under, where is actions that are permitted (but not obliged) under. = if is prohibited under 13
14 Behavior of an Obfuscator : Assumptions regarding beliefs This study: Supervisor and agent share knowledge regarding - (Choice) set of actions - Set of rules - Scores of actions on rules as in Future work: Diverging ideas about,, This study: Supervisor has no idea beforehand about which rule is used by the agent ( uninformative priors ) Future work: Repeated choices, iterative learning 14
15 Behavior of an Obfuscator Agent only cares about making sure that the supervisor remains unable to infer what rule lead to agent s choice. That is: agent has no motivation, other than obfuscation Easily extended to hybrid version, where rule-following and Obfuscation (measured by ) both play a role (more realistic) = But for ease of illustration, only two extreme cases are discussed in the remainder: Rule-followers and Obfuscators 15
16 Intermezzo: learning vs finding evidence Notion of a supervisor learning an agent s rules has no meaning when the agent only cares about obfuscation (and as such strictly speaking has no rules) Notion of finding evidence that a certain action follows from a certain rule is still meaningful in that context. Supervisor wishes to link an action to a rule beyond reasonable doubt Agent wishes to create reasonable doubt (measured in Entropy) 16
17 Behavior of an Obfuscator (II) Agent knows that supervisor will update his beliefs regarding the rule presumably used by her decision making: This posterior can be modeled as follows (Bayes, 1763): = prior: 1 Where: = 1 if is obliged under strong rule = if is prohibited under = 1 if is permitted under weak rule 17
18 Behavior of an Obfuscator (III) The agent knows that the supervisor s updated beliefs after having witnessed action result in uncertainty (denoted ). How to model, quantify this uncertainty? Using the notion of information Entropy (Shannon, 1948): = $ log Agent s behavior characterized by: argmax.. 18
19 Obfuscator worked out example -. - / / e.g. -. obliged by 1., permitted by 1 /, prohibited by 1, 1 2. e.g. 1. obliges -., prohibits - /, -. 19
20 Obfuscator worked out example (II) Rule-posteriors conditional on choosing action 1, 2, 3 P a r = 1; P a r 4 =.5; P a r 6 = ; P a r 7 = = = / = = = = P a 4 r = ; P a 4 r 4 =.5; P a 4 r 6 =.5; P a 4 r 7 = / = ; 9 1 / - / = / = 1 4 ; / = = 9 1 / - = = ; = 1 2
21 Obfuscator worked out example (III) Entropy resulting from choosing action 1: = 2 3 log log 1 3 =.276 Entropy resulting from choosing action 2: 4 = 1 4 log log log 1 2 =.452 Entropy resulting from choosing action 3: 6 = 1 log 1 = 21
22 Intermezzo behavior of Hybrid -. - / / Obfuscator will prefer 4 over, 6. Hybrid following 4 will prefer 4 over Hybrid following 1 will prefer - / over - Hybrid following may even prefer 4 over, depending on 22
23 Intermezzo: Maximizing entropy Maximizing rule-compliances / / =.24 4 =.29 6 =.29 Even though all actions are permitted by two rules, making all actions equivalent to a rule-compliance-maximizer, - / and - are preferred over -. by an Obfuscator. 23
24 Model extensions 24
25 Considered extensions From passive to active supervisor choice set design Versatile agent multiple rules 25
26 Behavior of active supervisor Until now: (implicit) assumption of passive supervisor : only exists in mind of the agent Now: supervisor is able to determine the set (A) of actions from which the agent chooses. - Select a set A of a given size - Given set of a given size, select alternative to add - Given set of a given size, select alternative to remove Supervisor s objective: minimize entropy contained in set Caveat: Entropy of action is contingent on set 26
27 Intermezzo: Entropy of action is contingent on set example -. - / / B -/ -.,- / = C. 2E B -/ - /,- = C. 2D 27
28 Behavior of active supervisor (II) Objective: minimize entropy through choice set design Depends on whether or not supervisor is naive or cynical Naive supervisor: Assumes agent follows some rule, does not care about obfuscating Cynical supervisor : Assumes agent only cares about obfuscating, has no rules Hybrid version covered in the paper Disclaimer: strictly speaking, naive and cynical not right terms 28
29 Behavior of naive supervisor Cf. efficient experimental design of DCEs Objective: minimize entropy through choice set design Is a function of the alternative chosen by the agent from the set; which depends on rule followed by the agent; which is unknown to the supervisor a priori. This is captured as follows: supervisor composes A such that: $ G $ G < $ G I $ G I K L Expected entropy associated with A Expected entropy associated with A 29
30 Behavior of naive supervisor Objective: minimize entropy through choice set design Is a function of the alternative chosen by the agent from the set; which depends on rule followed by the agent; which is unknown to the supervisor a priori. This is captured as follows: supervisor composes A such that: $ G $ G < $ G I $ G I K L Entropy associated with agent choosing action from A 3
31 Behavior of naive supervisor Objective: minimize entropy through choice set design Is a function of the alternative chosen by the agent from the set; which depends on rule followed by the agent; which is unknown to the supervisor a priori. This is captured as follows: supervisor composes A such that: $ G $ G < $ G I $ G I K L Probability of agent choosing action from A 31
32 Behavior of naive supervisor Objective: minimize entropy through choice set design Is a function of the alternative chosen by the agent from the set; which depends on rule followed by the agent; which is unknown to the supervisor a priori. This is captured as follows: supervisor composes A such that: $ G $ G < $ G I $ G I K L Probability of agent following rule multiplied by probability of choosing from K conditional Regret in Traveler Decision on following Making 32
33 Naive supervisor simple example -. - / / 1 N, 4 =.4 N 4, 6 =.38 P B -., - = C. /C $ G $ G = =.4 33
34 Behavior of cynical supervisor Objective: minimize entropy through choice set design Is a function of the alternative chosen by the agent from the set; which is based on obfuscation by the agent. This is captured as follows: supervisor composes A such that: max.. G < max.. G I K L Entropy of the maximum entropy action in A Entropy of the maximum entropy action in A 34
35 Cynical supervisor simple example -. - / / 1, 4 =.45; 4, 6 =.46; B -., - = C. C, 4, 4, 6 change order, compared to naive supervisor 35
36 Agent- supervisor interaction: Stackelberg game Supervisor = Leader: selects choice set, anticipating behavior of follower (agent) Agent = Follower: chooses action, given choice set provided by leader (supervisor) Resulting B Rule-follower agent Obfuscator agent Naive supervisor B QR B QRS Cynical supervisor B AR B ARS All resulting (expected) entropies can be computed using equations presented on earlier slides. 36
37 Considered extensions From passive to active supervisor choice set design Versatile agent multiple rules 37
38 Behavior of a versatile agent: Case of Rule-follower Agent cares about multiple rules, to different degrees. Denote T the utility (to the agent) of complying with the score on rule of action. Strong rule, normalized: T U R, Weak rule, normalized: T U R, Weight U R or shorthand U represents the relative importance of rule to the agent. Then, under simple MNL-assumptions: -dimensional vector of U s V = exp T exp T Regret in Traveler Decision Making 38
39 Behavior of a versatile agent: Case of Obfuscator Agent cares about prohibiting the supervisor to learn V. Maximize = Y Z V log Z V [V \ Where: Z V = ] ^_ V ` V Y ] ^_ V ` V av b Here: V = cde h gij f _g k _ij h gij cde f _g Modeled by means of continuous versions of Bayes theorem, Shannon entropy Note: for versatile Hybrid : = 1 T + 39
40 (much) Work to be done 1. Empirical: formulate choice models, estimate, validate. Study identifiability of parameters (e.g. ), perform DCEs 2. Theoretical: relax assumptions (e.g. regarding beliefs), study model properties, also as Multi-agent-systems 3. Simulations: norm formation in societies (infering norms from actions) using Agent-based models 4. Implications: different agents (humans, state actors, AIs), different contexts (military, health, transport, ) 4
Why on earth did you do that?!
Why on earth did you do that?! A simple model of obfuscation-based choice Caspar 9-5-2018 Chorus http://behave.tbm.tudelft.nl/ Delft University of Technology Challenge the future Background: models of
More informationLearning Equilibrium as a Generalization of Learning to Optimize
Learning Equilibrium as a Generalization of Learning to Optimize Dov Monderer and Moshe Tennenholtz Faculty of Industrial Engineering and Management Technion Israel Institute of Technology Haifa 32000,
More information9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering
Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make
More informationIntroduction to Game Theory
COMP323 Introduction to Computational Game Theory Introduction to Game Theory Paul G. Spirakis Department of Computer Science University of Liverpool Paul G. Spirakis (U. Liverpool) Introduction to Game
More informationRationality and Uncertainty
Rationality and Uncertainty Based on papers by Itzhak Gilboa, Massimo Marinacci, Andy Postlewaite, and David Schmeidler Warwick Aug 23, 2013 Risk and Uncertainty Dual use of probability: empirical frequencies
More informationOn the errors introduced by the naive Bayes independence assumption
On the errors introduced by the naive Bayes independence assumption Author Matthijs de Wachter 3671100 Utrecht University Master Thesis Artificial Intelligence Supervisor Dr. Silja Renooij Department of
More informationCS 570: Machine Learning Seminar. Fall 2016
CS 570: Machine Learning Seminar Fall 2016 Class Information Class web page: http://web.cecs.pdx.edu/~mm/mlseminar2016-2017/fall2016/ Class mailing list: cs570@cs.pdx.edu My office hours: T,Th, 2-3pm or
More informationTHE EMPIRICAL IMPLICATIONS OF PRIVACY-AWARE CHOICE
THE EMPIRICAL IMPLICATIONS OF PRIVACY-AWARE CHOICE RACHEL CUMMINGS, FEDERICO ECHENIQUE, AND ADAM WIERMAN Abstract. This paper initiates the study of the testable implications of choice data in settings
More informationGame-Theoretic Foundations for Norms
Game-Theoretic Foundations for Norms Guido Boella Dipartimento di Informatica Università di Torino-Italy E-mail: guido@di.unito.it Leendert van der Torre Department of Computer Science University of Luxembourg
More informationCOMP3702/7702 Artificial Intelligence Week1: Introduction Russell & Norvig ch.1-2.3, Hanna Kurniawati
COMP3702/7702 Artificial Intelligence Week1: Introduction Russell & Norvig ch.1-2.3, 3.1-3.3 Hanna Kurniawati Today } What is Artificial Intelligence? } Better know what it is first before committing the
More informationCS 188: Artificial Intelligence Spring Today
CS 188: Artificial Intelligence Spring 2006 Lecture 9: Naïve Bayes 2/14/2006 Dan Klein UC Berkeley Many slides from either Stuart Russell or Andrew Moore Bayes rule Today Expectations and utilities Naïve
More informationCoherence with Proper Scoring Rules
Coherence with Proper Scoring Rules Mark Schervish, Teddy Seidenfeld, and Joseph (Jay) Kadane Mark Schervish Joseph ( Jay ) Kadane Coherence with Proper Scoring Rules ILC, Sun Yat-Sen University June 2010
More informationMachine Learning for Large-Scale Data Analysis and Decision Making A. Week #1
Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Week #1 Today Introduction to machine learning The course (syllabus) Math review (probability + linear algebra) The future
More informationOther-Regarding Preferences: Theory and Evidence
Other-Regarding Preferences: Theory and Evidence June 9, 2009 GENERAL OUTLINE Economic Rationality is Individual Optimization and Group Equilibrium Narrow version: Restrictive Assumptions about Objective
More informationTheory Field Examination Game Theory (209A) Jan Question 1 (duopoly games with imperfect information)
Theory Field Examination Game Theory (209A) Jan 200 Good luck!!! Question (duopoly games with imperfect information) Consider a duopoly game in which the inverse demand function is linear where it is positive
More informationBelief and Desire: On Information and its Value
Belief and Desire: On Information and its Value Ariel Caticha Department of Physics University at Albany SUNY ariel@albany.edu Info-Metrics Institute 04/26/2013 1 Part 1: Belief 2 What is information?
More informationMachine Learning for Biomedical Engineering. Enrico Grisan
Machine Learning for Biomedical Engineering Enrico Grisan enrico.grisan@dei.unipd.it Curse of dimensionality Why are more features bad? Redundant features (useless or confounding) Hard to interpret and
More informationInformation, Utility & Bounded Rationality
Information, Utility & Bounded Rationality Pedro A. Ortega and Daniel A. Braun Department of Engineering, University of Cambridge Trumpington Street, Cambridge, CB2 PZ, UK {dab54,pao32}@cam.ac.uk Abstract.
More informationCorrelated Equilibrium in Games with Incomplete Information
Correlated Equilibrium in Games with Incomplete Information Dirk Bergemann and Stephen Morris Econometric Society Summer Meeting June 2012 Robust Predictions Agenda game theoretic predictions are very
More informationEvidence with Uncertain Likelihoods
Evidence with Uncertain Likelihoods Joseph Y. Halpern Cornell University Ithaca, NY 14853 USA halpern@cs.cornell.edu Riccardo Pucella Cornell University Ithaca, NY 14853 USA riccardo@cs.cornell.edu Abstract
More informationArea I: Contract Theory Question (Econ 206)
Theory Field Exam Winter 2011 Instructions You must complete two of the three areas (the areas being (I) contract theory, (II) game theory, and (III) psychology & economics). Be sure to indicate clearly
More informationToday s Outline. Recap: MDPs. Bellman Equations. Q-Value Iteration. Bellman Backup 5/7/2012. CSE 473: Artificial Intelligence Reinforcement Learning
CSE 473: Artificial Intelligence Reinforcement Learning Dan Weld Today s Outline Reinforcement Learning Q-value iteration Q-learning Exploration / exploitation Linear function approximation Many slides
More informationDiscrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10
EECS 70 Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 Introduction to Basic Discrete Probability In the last note we considered the probabilistic experiment where we flipped
More informationGuilt in Games. P. Battigalli and M. Dufwenberg (AER, 2007) Presented by Luca Ferocino. March 21 st,2014
Guilt in Games P. Battigalli and M. Dufwenberg (AER, 2007) Presented by Luca Ferocino March 21 st,2014 P. Battigalli and M. Dufwenberg (AER, 2007) Guilt in Games 1 / 29 ADefinitionofGuilt Guilt is a cognitive
More informationClustering, K-Means, EM Tutorial
Clustering, K-Means, EM Tutorial Kamyar Ghasemipour Parts taken from Shikhar Sharma, Wenjie Luo, and Boris Ivanovic s tutorial slides, as well as lecture notes Organization: Clustering Motivation K-Means
More informationShould all Machine Learning be Bayesian? Should all Bayesian models be non-parametric?
Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric? Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/
More informationBayesian Persuasion Online Appendix
Bayesian Persuasion Online Appendix Emir Kamenica and Matthew Gentzkow University of Chicago June 2010 1 Persuasion mechanisms In this paper we study a particular game where Sender chooses a signal π whose
More informationBasics of reinforcement learning
Basics of reinforcement learning Lucian Buşoniu TMLSS, 20 July 2018 Main idea of reinforcement learning (RL) Learn a sequential decision policy to optimize the cumulative performance of an unknown system
More informationGenerative Adversarial Networks (GANs) Ian Goodfellow, OpenAI Research Scientist Presentation at Berkeley Artificial Intelligence Lab,
Generative Adversarial Networks (GANs) Ian Goodfellow, OpenAI Research Scientist Presentation at Berkeley Artificial Intelligence Lab, 2016-08-31 Generative Modeling Density estimation Sample generation
More informationIntrinsic and Extrinsic Motivation
Intrinsic and Extrinsic Motivation Roland Bénabou Jean Tirole. Review of Economic Studies 2003 Bénabou and Tirole Intrinsic and Extrinsic Motivation 1 / 30 Motivation Should a child be rewarded for passing
More informationReinforcement Learning Wrap-up
Reinforcement Learning Wrap-up Slides courtesy of Dan Klein and Pieter Abbeel University of California, Berkeley [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationDefinitions and Proofs
Giving Advice vs. Making Decisions: Transparency, Information, and Delegation Online Appendix A Definitions and Proofs A. The Informational Environment The set of states of nature is denoted by = [, ],
More informationIntegrating State Constraints and Obligations in Situation Calculus
Integrating State Constraints and Obligations in Situation Calculus Robert Demolombe ONERA-Toulouse 2, Avenue Edouard Belin BP 4025, 31055 Toulouse Cedex 4, France. Robert.Demolombe@cert.fr Pilar Pozos
More informationGreat Expectations. Part I: On the Customizability of Generalized Expected Utility*
Great Expectations. Part I: On the Customizability of Generalized Expected Utility* Francis C. Chu and Joseph Y. Halpern Department of Computer Science Cornell University Ithaca, NY 14853, U.S.A. Email:
More informationPreferences for Randomization and Anticipation
Preferences for Randomization and Anticipation Yosuke Hashidate Abstract In decision theory, it is not generally postulated that decision makers randomize their choices. In contrast, in real life, even
More informationCS 188: Artificial Intelligence
CS 188: Artificial Intelligence Adversarial Search II Instructor: Anca Dragan University of California, Berkeley [These slides adapted from Dan Klein and Pieter Abbeel] Minimax Example 3 12 8 2 4 6 14
More informationShort Course: Multiagent Systems. Multiagent Systems. Lecture 1: Basics Agents Environments. Reinforcement Learning. This course is about:
Short Course: Multiagent Systems Lecture 1: Basics Agents Environments Reinforcement Learning Multiagent Systems This course is about: Agents: Sensing, reasoning, acting Multiagent Systems: Distributed
More information12 : Variational Inference I
10-708: Probabilistic Graphical Models, Spring 2015 12 : Variational Inference I Lecturer: Eric P. Xing Scribes: Fattaneh Jabbari, Eric Lei, Evan Shapiro 1 Introduction Probabilistic inference is one of
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 12: Probability 3/2/2011 Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Announcements P3 due on Monday (3/7) at 4:59pm W3 going out
More informationQuantum Games. Quantum Strategies in Classical Games. Presented by Yaniv Carmeli
Quantum Games Quantum Strategies in Classical Games Presented by Yaniv Carmeli 1 Talk Outline Introduction Game Theory Why quantum games? PQ Games PQ penny flip 2x2 Games Quantum strategies 2 Game Theory
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Introduction. Basic Probability and Bayes Volkan Cevher, Matthias Seeger Ecole Polytechnique Fédérale de Lausanne 26/9/2011 (EPFL) Graphical Models 26/9/2011 1 / 28 Outline
More informationProbability theory basics
Probability theory basics Michael Franke Basics of probability theory: axiomatic definition, interpretation, joint distributions, marginalization, conditional probability & Bayes rule. Random variables:
More informationGame Theory for Linguists
Fritz Hamm, Roland Mühlenbernd 4. Mai 2016 Overview Overview 1. Exercises 2. Contribution to a Public Good 3. Dominated Actions Exercises Exercise I Exercise Find the player s best response functions in
More informationMachine Learning for NLP
Machine Learning for NLP Linear Models Joakim Nivre Uppsala University Department of Linguistics and Philology Slides adapted from Ryan McDonald, Google Research Machine Learning for NLP 1(26) Outline
More informationReinforcement Learning
Reinforcement Learning Model-Based Reinforcement Learning Model-based, PAC-MDP, sample complexity, exploration/exploitation, RMAX, E3, Bayes-optimal, Bayesian RL, model learning Vien Ngo MLR, University
More informationBayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction
15-0: Learning vs. Deduction Artificial Intelligence Programming Bayesian Learning Chris Brooks Department of Computer Science University of San Francisco So far, we ve seen two types of reasoning: Deductive
More informationCS 188: Artificial Intelligence. Outline
CS 188: Artificial Intelligence Lecture 21: Perceptrons Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. Outline Generative vs. Discriminative Binary Linear Classifiers Perceptron Multi-class
More informationTowards a General Theory of Non-Cooperative Computation
Towards a General Theory of Non-Cooperative Computation (Extended Abstract) Robert McGrew, Ryan Porter, and Yoav Shoham Stanford University {bmcgrew,rwporter,shoham}@cs.stanford.edu Abstract We generalize
More informationLecture 16 Deep Neural Generative Models
Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed
More information13 : Variational Inference: Loopy Belief Propagation and Mean Field
10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction
More informationMachine Learning 3. week
Machine Learning 3. week Entropy Decision Trees ID3 C4.5 Classification and Regression Trees (CART) 1 What is Decision Tree As a short description, decision tree is a data classification procedure which
More informationCOMP3702/7702 Artificial Intelligence Lecture 11: Introduction to Machine Learning and Reinforcement Learning. Hanna Kurniawati
COMP3702/7702 Artificial Intelligence Lecture 11: Introduction to Machine Learning and Reinforcement Learning Hanna Kurniawati Today } What is machine learning? } Where is it used? } Types of machine learning
More informationSelecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden
1 Selecting Efficient Correlated Equilibria Through Distributed Learning Jason R. Marden Abstract A learning rule is completely uncoupled if each player s behavior is conditioned only on his own realized
More informationMarks. bonus points. } Assignment 1: Should be out this weekend. } Mid-term: Before the last lecture. } Mid-term deferred exam:
Marks } Assignment 1: Should be out this weekend } All are marked, I m trying to tally them and perhaps add bonus points } Mid-term: Before the last lecture } Mid-term deferred exam: } This Saturday, 9am-10.30am,
More informationDiscrete Mathematics and Probability Theory Fall 2015 Lecture 21
CS 70 Discrete Mathematics and Probability Theory Fall 205 Lecture 2 Inference In this note we revisit the problem of inference: Given some data or observations from the world, what can we infer about
More informationModels of Reputation with Bayesian Updating
Models of Reputation with Bayesian Updating Jia Chen 1 The Tariff Game (Downs and Rocke 1996) 1.1 Basic Setting Two states, A and B, are setting the tariffs for trade. The basic setting of the game resembles
More information15-381: Artificial Intelligence. Decision trees
15-381: Artificial Intelligence Decision trees Bayes classifiers find the label that maximizes: Naïve Bayes models assume independence of the features given the label leading to the following over documents
More informationStructure learning in human causal induction
Structure learning in human causal induction Joshua B. Tenenbaum & Thomas L. Griffiths Department of Psychology Stanford University, Stanford, CA 94305 jbt,gruffydd @psych.stanford.edu Abstract We use
More informationNotes on Mechanism Designy
Notes on Mechanism Designy ECON 20B - Game Theory Guillermo Ordoñez UCLA February 0, 2006 Mechanism Design. Informal discussion. Mechanisms are particular types of games of incomplete (or asymmetric) information
More informationLearning from Data. Amos Storkey, School of Informatics. Semester 1. amos/lfd/
Semester 1 http://www.anc.ed.ac.uk/ amos/lfd/ Introduction Welcome Administration Online notes Books: See website Assignments Tutorials Exams Acknowledgement: I would like to that David Barber and Chris
More informationEC476 Contracts and Organizations, Part III: Lecture 2
EC476 Contracts and Organizations, Part III: Lecture 2 Leonardo Felli 32L.G.06 19 January 2015 Moral Hazard: Consider the contractual relationship between two agents (a principal and an agent) The principal
More informationCOMP 551 Applied Machine Learning Lecture 19: Bayesian Inference
COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551 Unless otherwise noted, all material posted
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationBayesian networks in Mastermind
Bayesian networks in Mastermind Jiří Vomlel http://www.utia.cas.cz/vomlel/ Laboratory for Intelligent Systems Inst. of Inf. Theory and Automation University of Economics Academy of Sciences Ekonomická
More informationHow generative models develop in predictive processing
Faculty of Social Sciences Bachelor Artificial Intelligence Academic year 2016-2017 Date: 18 June 2017 How generative models develop in predictive processing Bachelor s Thesis Artificial Intelligence Author:
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationBargaining, Information Networks and Interstate
Bargaining, Information Networks and Interstate Conflict Erik Gartzke Oliver Westerwinter UC, San Diego Department of Political Sciene egartzke@ucsd.edu European University Institute Department of Political
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationThis corresponds to a within-subject experiment: see same subject make choices from different menus.
Testing Revealed Preference Theory, I: Methodology The revealed preference theory developed last time applied to a single agent. This corresponds to a within-subject experiment: see same subject make choices
More informationCommunication with Self-Interested Experts Part II: Models of Cheap Talk
Communication with Self-Interested Experts Part II: Models of Cheap Talk Margaret Meyer Nuffield College, Oxford 2013 Cheap Talk Models 1 / 27 Setting: Decision-maker (P) receives advice from an advisor
More informationStrategies under Strategic Uncertainty
Discussion Paper No. 18-055 Strategies under Strategic Uncertainty Helene Mass Discussion Paper No. 18-055 Strategies under Strategic Uncertainty Helene Mass Download this ZEW Discussion Paper from our
More informationOn the Unique D1 Equilibrium in the Stackelberg Model with Asymmetric Information Janssen, M.C.W.; Maasland, E.
Tilburg University On the Unique D1 Equilibrium in the Stackelberg Model with Asymmetric Information Janssen, M.C.W.; Maasland, E. Publication date: 1997 Link to publication General rights Copyright and
More informationGame Theory. Professor Peter Cramton Economics 300
Game Theory Professor Peter Cramton Economics 300 Definition Game theory is the study of mathematical models of conflict and cooperation between intelligent and rational decision makers. Rational: each
More informationMulticlass Classification-1
CS 446 Machine Learning Fall 2016 Oct 27, 2016 Multiclass Classification Professor: Dan Roth Scribe: C. Cheng Overview Binary to multiclass Multiclass SVM Constraint classification 1 Introduction Multiclass
More informationMarkov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graph model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationOverview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated
Fall 3 Computer Vision Overview of Statistical Tools Statistical Inference Haibin Ling Observation inference Decision Prior knowledge http://www.dabi.temple.edu/~hbling/teaching/3f_5543/index.html Bayesian
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationA survey: The convex optimization approach to regret minimization
A survey: The convex optimization approach to regret minimization Elad Hazan September 10, 2009 WORKING DRAFT Abstract A well studied and general setting for prediction and decision making is regret minimization
More informationQuestion 1. (p p) (x(p, w ) x(p, w)) 0. with strict inequality if x(p, w) x(p, w ).
University of California, Davis Date: August 24, 2017 Department of Economics Time: 5 hours Microeconomics Reading Time: 20 minutes PRELIMINARY EXAMINATION FOR THE Ph.D. DEGREE Please answer any three
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationDragon-kings. Quantum Decision Theory with Prospect Interference and Entanglement. Financial Crisis Observatory.
Quantum Decision Theory with Prospect Interference and Entanglement Didier Sornette (ETH Zurich) Dragon-kings (with V.I. Yukalov + PhD M. Favre and T. Kovalenko) Professor of Entrepreneurial Risks at ETH
More informationCollaborative topic models: motivations cont
Collaborative topic models: motivations cont Two topics: machine learning social network analysis Two people: " boy Two articles: article A! girl article B Preferences: The boy likes A and B --- no problem.
More informationCSE250A Fall 12: Discussion Week 9
CSE250A Fall 12: Discussion Week 9 Aditya Menon (akmenon@ucsd.edu) December 4, 2012 1 Schedule for today Recap of Markov Decision Processes. Examples: slot machines and maze traversal. Planning and learning.
More informationLearning Tetris. 1 Tetris. February 3, 2009
Learning Tetris Matt Zucker Andrew Maas February 3, 2009 1 Tetris The Tetris game has been used as a benchmark for Machine Learning tasks because its large state space (over 2 200 cell configurations are
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More informationBayesian consistent prior selection
Bayesian consistent prior selection Christopher P. Chambers and Takashi Hayashi August 2005 Abstract A subjective expected utility agent is given information about the state of the world in the form of
More informationReinforcement Learning Active Learning
Reinforcement Learning Active Learning Alan Fern * Based in part on slides by Daniel Weld 1 Active Reinforcement Learning So far, we ve assumed agent has a policy We just learned how good it is Now, suppose
More informationCostly Social Learning and Rational Inattention
Costly Social Learning and Rational Inattention Srijita Ghosh Dept. of Economics, NYU September 19, 2016 Abstract We consider a rationally inattentive agent with Shannon s relative entropy cost function.
More informationReview of Basic Probability
Review of Basic Probability Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 01003 September 16, 2009 Abstract This document reviews basic discrete
More informationEconomics 3012 Strategic Behavior Andy McLennan October 20, 2006
Economics 301 Strategic Behavior Andy McLennan October 0, 006 Lecture 11 Topics Problem Set 9 Extensive Games of Imperfect Information An Example General Description Strategies and Nash Equilibrium Beliefs
More informationAdvanced Topics in Machine Learning and Algorithmic Game Theory Fall semester, 2011/12
Advanced Topics in Machine Learning and Algorithmic Game Theory Fall semester, 2011/12 Lecture 4: Multiarmed Bandit in the Adversarial Model Lecturer: Yishay Mansour Scribe: Shai Vardi 4.1 Lecture Overview
More informationBayes Correlated Equilibrium and Comparing Information Structures
Bayes Correlated Equilibrium and Comparing Information Structures Dirk Bergemann and Stephen Morris Spring 2013: 521 B Introduction game theoretic predictions are very sensitive to "information structure"
More informationRobust Mechanism Design and Robust Implementation
Robust Mechanism Design and Robust Implementation joint work with Stephen Morris August 2009 Barcelona Introduction mechanism design and implementation literatures are theoretical successes mechanisms
More informationCS540 ANSWER SHEET
CS540 ANSWER SHEET Name Email 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 1 2 Final Examination CS540-1: Introduction to Artificial Intelligence Fall 2016 20 questions, 5 points
More informationLogistic Regression. Machine Learning Fall 2018
Logistic Regression Machine Learning Fall 2018 1 Where are e? We have seen the folloing ideas Linear models Learning as loss minimization Bayesian learning criteria (MAP and MLE estimation) The Naïve Bayes
More informationQuantifying information and uncertainty
Quantifying information and uncertainty Alexander Frankel and Emir Kamenica University of Chicago March 2018 Abstract We examine ways to measure the amount of information generated by a piece of news and
More informationQuantifying information and uncertainty
Quantifying information and uncertainty Alexander Frankel and Emir Kamenica University of Chicago April 2018 Abstract We examine ways to measure the amount of information generated by a piece of news and
More information1 [15 points] Search Strategies
Probabilistic Foundations of Artificial Intelligence Final Exam Date: 29 January 2013 Time limit: 120 minutes Number of pages: 12 You can use the back of the pages if you run out of space. strictly forbidden.
More informationEvidence with Uncertain Likelihoods
Evidence with Uncertain Likelihoods Joseph Y. Halpern Cornell University Ithaca, NY 14853 USA halpern@cs.cornell.edu Riccardo ucella Northeastern University Boston, MA 2115 USA riccardo@ccs.neu.edu Abstract
More information