Introduction to Spring 2009 Artificial Intelligence Midterm Exam

Similar documents
Introduction to Spring 2009 Artificial Intelligence Midterm Solutions

CS221 Practice Midterm

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

Introduction to Fall 2009 Artificial Intelligence Final Exam

Introduction to Fall 2008 Artificial Intelligence Midterm Exam

Introduction to Fall 2011 Artificial Intelligence Final Exam

Final. CS 188 Fall Introduction to Artificial Intelligence

Midterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

Section Handout 5 Solutions

Name: UW CSE 473 Final Exam, Fall 2014

Introduction to Spring 2006 Artificial Intelligence Practice Final

CS 188 Fall Introduction to Artificial Intelligence Midterm 2

CS 188 Fall Introduction to Artificial Intelligence Midterm 1

Midterm. Introduction to Artificial Intelligence. CS 188 Summer You have approximately 2 hours and 50 minutes.

CSL302/612 Artificial Intelligence End-Semester Exam 120 Minutes

Midterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes.

Mock Exam Künstliche Intelligenz-1. Different problems test different skills and knowledge, so do not get stuck on one problem.

CS 188 Introduction to AI Fall 2005 Stuart Russell Final

ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

Final Exam December 12, 2017

Midterm 2 V1. Introduction to Artificial Intelligence. CS 188 Spring 2015

Final Exam December 12, 2017

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

1 [15 points] Search Strategies

Final. CS 188 Fall Introduction to Artificial Intelligence

ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

CS 4100 // artificial intelligence. Recap/midterm review!

CS360 Homework 12 Solution

Final exam of ECE 457 Applied Artificial Intelligence for the Spring term 2007.

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.

Machine Learning, Midterm Exam

Midterm II. Introduction to Artificial Intelligence. CS 188 Fall ˆ You have approximately 3 hours.

Midterm Examination CS540: Introduction to Artificial Intelligence

Machine Learning, Midterm Exam: Spring 2009 SOLUTION

Introduction to Artificial Intelligence Midterm 2. CS 188 Spring You have approximately 2 hours and 50 minutes.

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.

Introduction to Fall 2008 Artificial Intelligence Final Exam

CSE 546 Final Exam, Autumn 2013

Final exam of ECE 457 Applied Artificial Intelligence for the Fall term 2007.

Introduction to Machine Learning Midterm, Tues April 8

Tentamen TDDC17 Artificial Intelligence 20 August 2012 kl

CS221 Practice Midterm #2 Solutions

CS 188 Fall Introduction to Artificial Intelligence Midterm 2

Using first-order logic, formalize the following knowledge:

Machine Learning, Fall 2009: Midterm

FINAL: CS 6375 (Machine Learning) Fall 2014

CS 170 Algorithms Spring 2009 David Wagner Final

Final Exam, Fall 2002

Final. Introduction to Artificial Intelligence. CS 188 Summer 2014

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20.

Final Exam, Machine Learning, Spring 2009

Minimax strategies, alpha beta pruning. Lirong Xia

Reinforcement Learning and Control

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability

Two hours UNIVERSITY OF MANCHESTER SCHOOL OF COMPUTER SCIENCE. Date: Thursday 17th May 2018 Time: 09:45-11:45. Please answer all Questions.

Midterm. Introduction to Machine Learning. CS 189 Spring You have 1 hour 20 minutes for the exam.

Final Exam, Spring 2006

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

CITS4211 Mid-semester test 2011

Written examination: Solution suggestions TIN175/DIT411, Introduction to Artificial Intelligence

Final Examination. Adrian Georgi Josh Karen Lee Min Nikos Tina. There are 12 problems totaling 150 points. Total time is 170 minutes.

Final Examination CS 540-2: Introduction to Artificial Intelligence

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability

Final Examination CS540-2: Introduction to Artificial Intelligence

GSOE9210 Engineering Decisions

CS 662 Sample Midterm

Midterm: CS 6375 Spring 2018

Final Exam. CS 188 Fall Introduction to Artificial Intelligence

Andrew/CS ID: Midterm Solutions, Fall 2006

CSE 473: Artificial Intelligence Autumn Topics

Q1. [24 pts] The OMNIBUS

CS 7180: Behavioral Modeling and Decisionmaking

This question has three parts, each of which can be answered concisely, but be prepared to explain and justify your concise answer.

Today s Outline. Recap: MDPs. Bellman Equations. Q-Value Iteration. Bellman Backup 5/7/2012. CSE 473: Artificial Intelligence Reinforcement Learning

Examination Artificial Intelligence Module Intelligent Interaction Design December 2014

Summary. Agenda. Games. Intelligence opponents in game. Expected Value Expected Max Algorithm Minimax Algorithm Alpha-beta Pruning Simultaneous Game

Figure 1: Bayes Net. (a) (2 points) List all independence and conditional independence relationships implied by this Bayes net.

CS 188: Artificial Intelligence Spring Announcements

FINAL EXAM PRACTICE PROBLEMS CMSC 451 (Spring 2016)

Math 16 - Practice Final

Introduction to Arti Intelligence

Be able to define the following terms and answer basic questions about them:

ARTIFICIAL INTELLIGENCE. Reinforcement learning

MARKOV DECISION PROCESSES (MDP) AND REINFORCEMENT LEARNING (RL) Versione originale delle slide fornita dal Prof. Francesco Lo Presti

Deep Reinforcement Learning. STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 19, 2017

Machine Learning Practice Page 2 of 2 10/28/13

Homework 2: MDPs and Search

Decision making, Markov decision processes

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

Learning Tetris. 1 Tetris. February 3, 2009

CSE 473: Artificial Intelligence Spring 2014

Introduction to Reinforcement Learning

Final. Introduction to Artificial Intelligence. CS 188 Fall You have approximately 2 hours and 50 minutes.

Physics 200 Exam 1 Kinematics and Force September 27, 2007

Transcription:

S 188 Introduction to Spring 009 rtificial Intelligence Midterm Exam INSTRUTINS You have 3 hours. The exam is closed book, closed notes except a one-page crib sheet. Please use non-programmable calculators only. Mark your answers N THE EM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation. ll short answer sections can be successfully answered in a few sentences at most. Last Name First Name SID Login GSI Section Time ll the work on this exam is my own. (please sign) For staff use only Q. 1 Q. Q. 3 Q. Q. 5 Q. 6 Total /16 /15 /0 /10 /1 /18 /100

1. (16 points) True/False For the following questions, a correct answer is worth points, no answer is worth 1 point, and an incorrect answer is worth 0 points. ircle true or false to indicate your answer. a) (true or false) If g(s) and h(s) are two admissible heuristics, then their average f(s) = 1 g(s) + 1 h(s) must also be admissible. b) (true or false) For a search problem, the path returned by uniform cost search may change if we add a positive constant to every step cost. c) (true or false) The running-time of an efficient solver for tree-structured constraint satisfaction problems is linear in the number of variables. d) (true or false) If h 1 (s) is a consistent heuristic and h (s) is an admissible heuristic, then min(h 1 (s), h (s)) must be consistent. e) (true or false) The amount of memory required to run minimax with alpha-beta pruning is (b d ) for branching factor b and depth limit d. f) (true or false) In a Markov decision process with discount γ = 1, the difference in values for two adjacent states is bounded by the reward between them: V (s) V (s ) max a R(s, a, s ). g) (true or false) Value iteration and policy iteration must always converge to the same policy. h) (true or false) In a Bayes net, if B, then B for some variable other than or B.

E! F! G NME: 3. (15 points) Search: rawler s Escape G G Whilst Pacman was Q-learning, rawler snuck into mediumlassic and stole all the dots. Now, it s trying to escape as quickly as possible. t each time step, G rawler can either move its shoulder G or its elbow one position up or down. Both joints s and e have five total positions each (1 through 5) and both begin in position 3. Upon changing arm positions from (s, e) to (s, e ), rawler s body moves some distance r(s, e, s, e ), where r(s, e, s, e ) meters (negative distance means the crawler moves backwards). rawler must travel 10 meters to reach the exit. N ction: increase elbow position exit r(3, 3, 3, ) 10 meters In this problem, you will design a search problem for which the optimal solution will allow rawler to escape in as few time steps as possible. (a) (3 pt) Define a state space, start state and goal test to represent this problem. (b) (3 pt) Define the successor function for the start state by listing its (successor state, action, step cost) triples. Use the actions s + and s for increasing and decreasing the shoulder position and e + and e for the elbow.

(c) ( pt) Would depth-first graph search be complete for the search problem you defined? Explain. (d) (3 pt) Design a non-trivial, admissible heuristic for this problem. (e) ( pt) rawler s shoulder overheats if it switches direction more than once in any three consecutive time steps, making progress impossible. Describe how you would change your search problem so that an optimal search algorithm would only return solutions that avoid overheating.

NME: 5 3. (0 points) SPs: onstraining Graph Structure Hired as a consultant, you built a Bayes net model of heeseboard Pizza customers. Last night, a corporate spy from Zachary s Pizza deleted the directions from 1! all your arcs.! You still have! the undirected graph of your E n! model and a list of dependencies. You now have to recover 1! E the direction! E of the arrows n-1! to build a graph that allows for these dependencies. Note: Y means is not independent of Y. E! B! 1!! B E n! 1! E! E n-1! Distribution properties (dependencies): G E! F! G F! G B F B G a) (1 pt) Given the first constraint only, circle all the topologies that are allowed for the triple (,, G). G G G G D P! T! P! N G G G G b) (3 pt) Formulate direction-recovery for this graph as a SP with explicit binary constraints only. The variables and values are provided for you. The variable E stands for the direction of the arc between and the center, where out means the arrow points outward (toward ), and in means the arrow points inward (toward N ). T! Variables: E, E B, E F, E G E! Values: out, in r(s 0, e 0, s 0, e 0 +1) 10 meters onstraints: exit exit E! r(s 0, e 0, s 0, e 0 +1) 10 meters c) (1 pt) Draw the constraint graph for this SP. d) ( pt) fter selecting E = in, cross out all values eliminated from E B, E F and E G by forward checking. E E B E F E G in out in out in out in e) ( pt) ross out all values eliminated by arc consistency applied before any backtracking search. E E B E F E G out in out in out in out in f) (1 pt) Solve this SP, then add the correct directions to the arcs of the graph at the top of the page.

6 Now consider a chain structured graph over variables 1,..., n. Edges E 1,..., E n 1 connect adjacent variables, but their directions are again unknown. B 1!!! E n! 1! E! E n-1! g) ( pt) Using only binary constraints and two-valued variables, formulate a SP that is satisfied by only and all networks that can represent a distribution where 1 n. Describe what your variables mean in terms of direction of the arrows in the network. Variables: D E! onstraints: E! E! E! E! h) (6 pt) Using only unary, binary and ternary (3 variable) constraints and two-valued variables, formulate a SP that is satisfied by only and all networks that enforce 1 n. Describe what your variables mean in terms of direction of the arrows in the network. Hint: You will have to introduce additional variables. Variables: onstraints:

NME: 7. (10 points) Multi-agent Search: onnect-3 In the game of onnect-3, players and alternate moves, dropping their symbols into columns 1,, 3, or. Three-in-a-row wins the game, horizontally, vertically or diagonally. plays first. 1 3 1 3 (a) (1 pt) What is the maximum branching factor of minimax search for onnect-3? (b) (1 pt) What is the maximum tree depth in plies? (c) (1 pt) Give a reasonably tight upper bound on the number of terminal states. (d) ( pt) Draw the game tree starting from the board shown above right (with to play next). You may 1 3 1 3 abbreviate states by drawing only the upper-right region circled. The root node is drawn for you. (e) ( pt) is the maximizer, while is the minimizer. s utility for terminal states is k when wins and k when wins, where k is 1 for a horizontal 3-in-a-row win (as in above left), for a vertical win, and 3 for a diagonal win. tie has value 0. Label each node of your tree with its minimax value. (f) (3 pt) ircle all nodes of your tree that will not be explored when using alpha-beta pruning and a move ordering that maximizes the number of nodes pruned.

8 5. (1 points) MDPs: Robot Soccer soccer robot is on a fast break toward the goal, starting in position 1. From positions 1 through 3, it can either shoot (S) or dribble the ball forward (D); from it can only shoot. If it shoots, it either scores a goal (state G) or misses (state M). If it dribbles, it either advances a square or loses the ball, ending up in M. 1 3 1 3 D Goal 1 3 In this MDP, the states are 1,, 3,, G and M, where G and M are terminal states. The transition model depends on the parameter y, which is the probability of dribbling success. ssume a discount of γ = 1. T (k, S, G) = k 6 T (k, S, M) = 1 k 6 for k {1,, 3, } T (k, D, k + 1) = y T (k, D, M) = 1 y for k {1,, 3} R(k, S, G) = 1 for k {1,, 3, }, and rewards are 0 for all other transitions (a) ( pt) What is V π (1) for the policy π that always shoots? (b) ( pt) What is Q (3, D) in terms of y? (c) ( pt) Using y = 3, complete the first two iterations of value iteration. i Vi (1) V i () V i (3) V i () 0 0 0 0 0 1 (d) ( pt) fter how many iterations will value iteration compute the optimal values for all states? (e) ( pt) For what range of values of y is Q (3, S) Q (3, D)?

NME: 9 The dribble success probability y in fact depend on the presence or absence of a defending robot, D. has no way of detecting whether D is present, but does know some statistical properties of its environment. D is present 3 of the time. When D is absent, y = 3. When D is present, y = 1. (f) ( pt) What is the posterior probability that D is present, given that Dribbles twice successfully from 1 to 3, then Shoots from state 3 and scores. (g) (3 pt) What transition model should use in order to correctly compute its maximum expected reward when it doesn t know whether or not D is present? T (k, S, G) = T (k, S, M) = for k {1,, 3, } T (k, D, k + 1) = T (k, D, M) = for k {1,, 3} (h) ( pt) What is the optimal policy π when doesn t know whether or not D is present?

B 10 D E! 6. (18 points) Bayes Nets: The Mind of Pacman Pacman doesn t just eat dots. He also enjoys pears, apples, carrots, nuts, eggs and toast. Each morning, he chooses to eat some subset of these foods. His preferences are modeled by a Bayes net with this structure. E! E P! T! N E! a) ( pt) Factor the probability that Pacman chooses to eat only apples and nuts in terms of conditional probabilities from this Bayes net. P ( p, a, c, n, e, t) = b) ( pt) For each of the following properties, circle whether they are true, false or unknown of the distribution P (P,,, N, E, T ) for this Bayes net. N T (true, false, unknown) P E N (true, false, unknown) P N, (true, false, unknown) E, N (true, false, unknown) c) ( pt) If the arrow from T to E were reversed, list two conditional independence properties that would be true under the new graph that were not guaranteed under the original graph. d) ( pt) You discover that P (a c, t) > P (a c). Suggest how you might change the network in response.

NME: 11 e) ( pt) You now want to compute the distribution P () using variable elimination. List the factors that remain before and after eliminating the variable N. Before: fter: 1!!! E n! 1! E! E n-1! B D E! f) ( pt) Pacman s new diet allows only fruit (P and ) to be eaten, but Pacman only follows the diet occasionally. dd the new variable D (for whether he follows the E! diet) to the network below E! by adding arcs. Briefly justify your answer. P! E! E! D T! E! N g) ( pt) Given the Bayes net and the factors below, fill in the table for P (D a) or state that there is not enough information. D P (D) d 0. d 0.6 D P ( D) d a 0.8 d a 0. d a 0.5 d a 0.5 D P (D = a) d a d a