Introduction to Spring 2009 Artificial Intelligence Midterm Solutions

Size: px
Start display at page:

Download "Introduction to Spring 2009 Artificial Intelligence Midterm Solutions"

Transcription

1 S 88 Introduction to Spring 009 rtificial Intelligence Midterm Solutions. (6 points) True/False For the following questions, a correct answer is worth points, no answer is worth point, and an incorrect answer is worth 0 points. ircle true or false to indicate your answer. a) (true or false) If g(s) and h(s) are two admissible heuristics, then their average f(s) = g(s) + h(s) must also be admissible. True. Let h (s) be the true distance from s. We know that g(s) h (s) and h(s) h (s), thus f(s) = g(s) + h(s) h (s) + h (s) = h (s) b) (true or false) For a search problem, the path returned by uniform cost search may change if we add a positive constant to every step cost. True. onsider that there are two paths from the start state (S) to the goal (G), S G and S G. cost(s, ) =, cost(, G) =, and cost(s, G) =. So the optimal path is through. Now, if we add to each of the costs, the optimal path is directly from S to G. Since uniform cost search finds the optimal path, its path will change. c) (true or false) The running-time of an efficient solver for tree-structured constraint satisfaction problems is linear in the number of variables. True. The running time of the algorithm for tree-structured SPs is (n d ), where n is the number of variables and d is the maximum size of any variable s domain. d) (true or false) If h (s) is a consistent heuristic and h (s) is an admissible heuristic, then min(h (s), h (s)) must be consistent. False. For instance, if h (s) be admissible but inconsistent, and h (s) dominate h (s), then min(h (s), h (s)) = h (s), which is inconsistent. e) (true or false) The amount of memory required to run minimax with alpha-beta pruning is (b d ) for branching factor b and depth limit d. True and False (everyone wins). The memory required is only (bd), so we accepted False. However, by definition an algorithm that is (bd) is also (b d ), because denotes upper bounds that may or may not be tight, so technically this statement is True (but not very useful). f) (true or false) In a Markov decision process with discount γ =, the difference in values for two adjacent states is bounded by the reward between them: V (s) V (s ) max a R(s, a, s ). False. Let V (s ) = 0, and R(s, a, s ) = 0 a, but there is an action a which takes s to the terminal state T and gives a reward 00. Thus V (s) 00, but the inequality above says that V (s) 0 0. g) (true or false) Value iteration and policy iteration must always converge to the same policy. True and False (everyone wins). Both algorithms are guaranteed to converge to the optimal policy, so we accepted True. If there are multiple policies that are optimal (meaning they yield the same maximal values for every state), then the algorithms might diverge. Both value iteration and policy iteration will always lead to the same optimal values. h) (true or false) In a Bayes net, if B, then B for some variable other than or B. False. onsider the Bayes net B. learly, B, but B.

2 E! F! G. (5 points) Search: rawler s Escape G G Whilst Pacman was Q-learning, rawler snuck into mediumlassic and stole all the dots. Now, it s trying to escape as quickly as possible. t each time step, G rawler can either move its shoulder G or its elbow one position up or down. Both joints s and e have five total positions each ( through 5) and both begin in position. Upon changing arm positions from (s, e) to (s, e ), rawler s body moves some distance r(s, e, s, e ), where r(s, e, s, e ) meters (negative distance means the crawler moves backwards). rawler must travel 0 meters to reach the exit. N ction: increase elbow position exit r(,,, ) 0 meters In this problem, you will design a search problem for which the optimal solution will allow rawler to escape in as few time steps as possible. (a) ( pt) Define a state space, start state and goal test to represent this problem. state is a -tuple consisting of distance to exit, shoulder position, and elbow position. The start state is (0,, ). Goal test: distance to exit is less than or equal to zero. (b) ( pt) Define the successor function for the start state by listing its (successor state, action, step cost) triples. Use the actions s + and s for increasing and decreasing the shoulder position and e + and e for the elbow. { ((0 r(,,, ),, ), s +, ), ((0 r(,,, ),, ), s, ), ((0 r(,,, ),, ), e +, ), ((0 r(,,, ),, ), e, ) }

3 NME: (c) ( pt) Would depth-first graph search be complete for the search problem you defined? Explain. No, depth-first search would not be complete because the state space is infinite. Depth-first search could forever keep expanding states with increasing distance to the goal. (d) ( pt) Design a non-trivial, admissible heuristic for this problem. Distance to exit divided by is admissible because each step can move rawler at most meters closer to the goal. slightly better heuristic would be to round this heuristic up to the nearest integer. (e) ( pt) rawler s shoulder overheats if it switches direction more than once in any three consecutive time steps, making progress impossible. Describe how you would change your search problem so that an optimal search algorithm would only return solutions that avoid overheating. onsider a reversal to be s + followed by s or s followed by s +. To avoid these, we can extend the state space to include the last two actions. We then remove from the output of the successor function any triple that includes the action sequences (s +, s, s + ) or (s, s +, s ). We also accepted solutions that gave a very high step cost to these successors.

4 . (0 points) SPs: onstraining Graph Structure Hired as a consultant, you built a Bayes net model of heeseboard Pizza customers. Last night, a corporate spy from Zachary s Pizza deleted the directions from! all your arcs.!! E You still have the undirected n! graph of your! E! E n-! model and a list of dependencies. You now have to recover the direction of the arrows to build a graph that allows for these dependencies. Note: Y means is not independent of Y. E! B Distribution properties (dependencies): G!!! n! E! E n-! B F! E! G F! G B F B G a) ( pt) Given the first constraint only, circle all the topologies that are allowed for the triple (,, G). G G G G P! T! P! G G r(s 0, e 0, s 0, e 0 +) G G b) ( pt) Formulate direction-recovery for this graph as a SP with explicit binary constraints only. The D variables and values are provided for you. The variable E stands for the direction of the arc between N ction: increase and the center, where out means the arrow points outward elbow position (toward ), and in means the arrow points T! inward (toward ). Variables: E, E B, E F, E G r(,,, ) Values: out, in 0 meters onstraints: n explicit constraint N lists all legal tuples of values are listed for a tuple of variables. (E, E G ) {(in, out), (out, out), (out, in)} (E B, E F ) {(in, out), (out, out), (out, in)} (E B, E G ) {(in, in)} exit 0 meters c) ( pt) Draw the constraint graph for this SP. E E G E B E F d) ( pt) fter selecting E = in, cross out all values eliminated from E B, E F and E G by forward checking. E E B E F E G in out in out in out in Forward checking removes the values of variables that directly contradict the assignment E = in. e) ( pt) ross out all values eliminated by arc consistency applied before any backtracking search. E E B E F E G out in out in out in out in We first remove E B = out because it is incompatible with any value for E G. Likewise, we remove E G = out. Now, we can remove E = in based on E G and E F = in based on E B. f) ( pt) Solve this SP, then add the correct directions to the arcs of the graph at the top of the page. E = out, E B = in, E F = out, and E G = in. exit

5 NME: 5 Now consider a chain structured graph over variables,..., n. Edges E,..., E n connect adjacent variables, but their directions are again unknown. B D!!! E n!! E! E n-! g) ( pt) Using only binary constraints and two-valued variables, formulate a SP that is satisfied by only and all networks that can represent a distribution where n. Describe what your variables mean in terms of direction of the arrows in the network. Variables: Variables E,..., E n correspond to the directions of these edges and take values {left, right}. onstraints: (E i, E i+ ) (right, left), for i {0,..., n }. h) (6 pt) Using only unary, binary andternary ( variable) constraints andtwo-valued variables, formulate a SP that is satisfied by only and all networks that enforce n. Describe what your variables mean in terms of direction of the arrows in the network. Hint: You will have to introduce additional variables. Variables: Variables E,..., E n correspond to the directions of these edges and take values {left, right}. I,..., I n can take values {T, F }.,..., n can take values {T, F }. onstraints: Intuitively, we want I i to be T if i i+ i+, i.e. the triple i, i+, i+ is inactive. lso, we want i to be T if either I i or i is T, i.e. if the current triple is inactive or if any of the prior triples was inactive. Finally, we want n to be T which would imply that there is some inactive triple. (I i, E i, E i+ ) {(T, right, left), (F, right, right), (F, left, right), (F, left, left)} for i {,..., n }. (I, ) {(T, T ), (F, F )} ( i, I i, i ) {(F, F, F ), (F, T, T ), (T, F, T ), (T, T, T )} for i {,..., n } n = true

6 6. (0 points) Multi-agent Search: onnect- In the game of onnect-, players and alternate moves, dropping their symbols into columns,,, or. Three-in-a-row wins the game, horizontally, vertically or diagonally. plays first. (a) ( pt) What is the maximum branching factor of minimax search for onnect-?. (b) ( pt) What is the maximum tree depth in plies? 6. ccording to our terminology, a search ply in a multi-agent search problem includes one move for every agent. Because the textbook differs in its definition, we also accepted. (c) ( pt) Give a reasonably tight upper bound on the number of terminal states. Several answers were accepted: is a reasonable bound on the number of leaves of the search tree. is a reasonable bound on the number of total states in the game, because each square can be empty, contain an or contain an. tighter bound is 5, which follows from the observation that there are only 5 ways to legally fill a column from the bottom up with s and s. Bounds based on the assumption that the entire board would be filled with s and s were not accepted, even if they in fact overestimated the true number of terminal states. (d) ( pt) Draw the game tree starting from the board shown above right (with to play next). You may abbreviate states by drawing only the upper-right region circled. The root node is drawn for you.!"#!"# $#!"# $#!%#!"#!%# (e) ( pt) is the maximizer, while is the minimizer. s utility for terminal states is k when wins and k when wins, where k is for a horizontal -in-a-row win (as in above left), for a vertical win, and for a diagonal win. tie has value 0. Label each node of your tree with its minimax value. (f) ( pt) ircle all nodes of your tree that will not be explored when using alpha-beta pruning and a move ordering that maximizes the number of nodes pruned.

7 NME: 7 5. ( points) MDPs: Robot Soccer soccer robot is on a fast break toward the goal, starting in position. From positions through, it can either shoot (S) or dribble the ball forward (D); from it can only shoot. If it shoots, it either scores a goal (state G) or misses (state M). If it dribbles, it either advances a square or loses the ball, ending up in M. D Goal In this MDP, the states are,,,, G and M, where G and M are terminal states. The transition model depends on the parameter y, which is the probability of dribbling success. ssume a discount of γ =. T (k, S, G) = k 6 T (k, S, M) = k 6 for k {,,, } T (k, D, k + ) = y T (k, D, M) = y for k {,, } R(k, S, G) = for k {,,, }, and rewards are 0 for all other transitions (a) ( pt) What is V π () for the policy π that always shoots? V π () = T (, S, G)R(, S, G) + T (, S, M)R(, S, M) = 6 (b) ( pt) What is Q (, D) in terms of y? Q (, D) = T (, D, )(R(, D, ) + V ()) + T (, D, M)R(, D, M) = T (, D, )V () = T (, D, )Q (, S) = T (, D, )(T (, S, G)R(, S, G) + T (, S, M)R(, S, m)) = T (, D, )T (, S, G)R(, S, G) = y (c) ( pt) Using y =, complete the first two iterations of value iteration. i Vi () V i () V i () V i () (d) ( pt) fter how many iterations will value iteration compute the optimal values for all states? fter iterations, the values will have converged when y =. bove, only V () has not yet converged. We note that for y >, a fourth iteration would be required because a fast break has up to four transitions. (e) ( pt) For what range of values of y is Q (, S) Q (, D)? Q (, S) Q (, D) T (, S, G) T (, D, ) T (, S, G) y y 0

8 8 The dribble success probability y in fact depend on the presence or absence of a defending robot, D. has no way of detecting whether D is present, but does know some statistical properties of its environment. D is present of the time. When D is absent, y =. When D is present, y =. (f) ( pt) What is the posterior probability that D is present, given that Dribbles twice successfully from to, then Shoots from state and scores. We can use Bayes rule, where D is a random variable denoting the presence of D, and e is the evidence that dribbled twice and scored. P (d e) = P (e d) P (d) P (e) P (e) = P (e d) P (d) + P (e d) P ( d) P (e d) = P (e d) = P (e) = + 9 = 96 P (d e) = 96 / 96 = (g) ( pt) What transition model should use in order to correctly compute its maximum expected reward when it doesn t know whether or not D is present? To maximize expected total reward, the agent should model the situation as accurately as possible. We accepted two answers for T (k, D, k + ). ne answer is to claim that the agent should use its marginal belief that each dribble will be successful, summing over the two possibilities of D s presence or absense. In this case, T (k, D, k + ) = + = 5 7, and T (k, D, M) =. nother acceptable answer would be to update the beliefs of the agent after every successful dribble. Hence, while T (, D, ) = 5 as above, reaching state implies a successful previous dribble, and so another successful dribble is more likely. In particular, let, and indicate successful first, second and third dribbles, while D is the presence of the defender. Since D, we have P (x x ) = d P (d x )P (x d). omputing similarly to part (f), P (d x ) = 5, so T (, D, ) = = 0. We can compute T (, D, ) similarly. is conditionally independent of both and given D, so P (x x, x ) = d P (d x, x )P (x d). Using our result from (f), P (d x, x ) =, and so T (, D, ) = + 9 = 9. Since a dribble either succeeds or fails, T (k, D, M) = T (k, D, k + ) for all k. Shooting probabilities are unchanged by the presence of the defender. T (k, S, G) = k 6 T (k, S, M) = k 6 for k {,,, } T (k, D, k + ) = 5 T (k, D, M) = 7 for k {,, } R T (k, S, G) = k 6 T (k, S, M) = k 6 for k {,,, } T (, D, ) = 5 T (, D, ) = 0 T (, D, ) = 9 T (, D, M) = 7 T (, D, M) = 9 0 T (, D, M) = 5

9 NME: 9 (h) ( pt) What is the optimal policy π when doesn t know whether or not D is present? Using T (k, D, k + ) = 5, we find that π = { : S, : S, : S, : S}. We showed from part (e) that for any y <, shooting is preferable to dribbling from state. Therefore, we know that V () = Q (, S). We can perform similar computations for states and : V () = Q (, S) = π () = S Q (, S) = Q (, D) = 5 V () = 5 V () = max(q (, S), Q (, D)) = π () = S Q (, S) = 6 Q (, D) = 5 V () = 5 6 V () = max(q (, S), Q (, D)) = 6 π () = S Under the second answer for part (g), similar computations give π = { : S, : S, : S, : S}. V () = Q (, S) = Q (, D) = 9 V () = V () = max(q (, S), Q (, D)) = π () = S Q (, S) = Q (, D) = 0 V () = V () = max(q (, S), Q (, D)) = π () = S Q (, S) = 6 Q (, D) = 5 V () = 5 6 V () = max(q (, S), Q (, D)) = 6 π () = S

10 B 0 D 6. (8 points) Bayes Nets: The Mind of Pacman Pacman doesn t just eat dots. He also enjoys pears, apples, carrots, nuts, eggs and toast. Each morning, he chooses to eat some subset of these foods. His preferences are modeled by a Bayes net with this structure. E P! T! N a) ( pt) Factor the probability that Pacman chooses to eat only apples and nuts in terms of conditional probabilities from this Bayes net. P ( p, a, c, n, e, t) = P (a) P ( p a) P ( c a) P (n p, c) P ( t c) P ( e n, t) b) ( pt) For each of the following properties, circle whether they are true, false or unknown of the distribution P (P,,, N, E, T ) for this Bayes net. Without the conditional probability tables, we do not know if any two variables are dependent. From the network topology, we can only conclude true or unknown for the questions below. N T (true, false, unknown) unknown: T N is an active path. P E N (true, false, unknown) unknown: P T E is an active path. P N, (true, false, unknown) unknown: P and N are directly connected. E, N (true, false, unknown) true: ll three paths from E to are inactive. c) ( pt) If the arrow from T to E were reversed, list two conditional independence properties that would be true under the new graph that were not guaranteed under the original graph. The triple T E is no longer active when T is unknown, therefore P E N and E N and E N The triple N E T is no longer active when E is known, therefore T N, E and T P, E and T, E There are further independence assumptions based on these 6 that condition on more variables, such as E N, P. d) ( pt) You discover that P (a c, t) > P (a c). Suggest how you might change the network in response. This inequality implies that T by definition, but in the network above, and T are conditionally independent given. So, the network must be changed to allow for this dependence. dding an arc from to T would fix the problem. Note: adding an arc from T to creates a cycle, which is not allowed in a Bayes net. ther acceptable solutions exist, like reversing the direction of the arc from T to.

11 NME: e) ( pt) You now want to compute the distribution P () using variable elimination. List the factors that remain before and after eliminating the variable N. Before: The initial factors are just the conditional probability tables of the Bayes net. P (), P (P ), P ( ), P (T ), P (N P, ), P (E T, N) fter: First, all factors that include the variable N are joined together, yielding P (E, N P,, T ) Next, the variable N is summed out of this new factor, yielding P (E P,, T ) The remaining factors include this new factor and the unused original factors. P (), P (P ), P ( ), P (T ), P (E P,, T ) Referring to the final factor as m(e, P,, T ) (like in the textbook) was also accepted. f) ( pt) Pacman s new diet allows only fruit (P and ) to be eaten, but Pacman only follows the diet occasionally. dd the new variable D (for whether he follows the diet) to the network below by adding arcs. Briefly justify your answer. P! P! D D T! N T! N The best answer is the one on the left, which allows us to express changes to all food preferences based on the diet. ther answers were accepted with appropriate justification, such as the network on the right, which specifies that the diet specifically disallows, N, E and T. onnecting D only to P and/or is problematic: for instance, observing P and would make D independent of, but the diet should still affect whether carrots () are allowed even when and P are known. g) ( pt) Given the Bayes net and the factors below, fill in the table for P (D a) or state that there is not enough information. From the tables P (D) and P ( D), the entries of the factor P (D = a) can be computed from Bayes rule (which always holds): no additional information about the properties of the distributions are required. The trick is to observe that while P ( a) is not explicitly given, it can be computed from P (D) and P ( D) using the product rule: a,d : P (a, d) = P (d)p (a d). omputing P ( a) in this way is equivalent to applying the normalization trick. D P (D) d 0. d 0.6 D P ( D) d a 0.8 d a 0. d a 0.5 d a 0.5 D P (D = a) P ( a d) P (d) d a P ( a d) P ( d)+p ( a d) P (d) = = 9 P ( a d) P ( d) d a P ( a d) P ( d)+p ( a d) P (d) = = 5 9

Introduction to Spring 2009 Artificial Intelligence Midterm Exam

Introduction to Spring 2009 Artificial Intelligence Midterm Exam S 188 Introduction to Spring 009 rtificial Intelligence Midterm Exam INSTRUTINS You have 3 hours. The exam is closed book, closed notes except a one-page crib sheet. Please use non-programmable calculators

More information

CS221 Practice Midterm

CS221 Practice Midterm CS221 Practice Midterm Autumn 2012 1 ther Midterms The following pages are excerpts from similar classes midterms. The content is similar to what we ve been covering this quarter, so that it should be

More information

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes. CS 188 Spring 2014 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Fall 2015 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Spring 2017 Introduction to Artificial Intelligence Midterm V2 You have approximately 80 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark

More information

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.

More information

Section Handout 5 Solutions

Section Handout 5 Solutions CS 188 Spring 2019 Section Handout 5 Solutions Approximate Q-Learning With feature vectors, we can treat values of states and q-states as linear value functions: V (s) = w 1 f 1 (s) + w 2 f 2 (s) +...

More information

Final. CS 188 Fall Introduction to Artificial Intelligence

Final. CS 188 Fall Introduction to Artificial Intelligence S 188 Fall 2012 Introduction to rtificial Intelligence Final You have approximately 3 hours. The exam is closed book, closed notes except your three one-page crib sheets. Please use non-programmable calculators

More information

Introduction to Fall 2011 Artificial Intelligence Final Exam

Introduction to Fall 2011 Artificial Intelligence Final Exam CS 188 Introduction to Fall 2011 rtificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except two pages of crib sheets. Please use non-programmable calculators

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Spring 2017 Introduction to Artificial Intelligence Midterm V2 You have approximately 80 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark

More information

Introduction to Fall 2008 Artificial Intelligence Midterm Exam

Introduction to Fall 2008 Artificial Intelligence Midterm Exam CS 188 Introduction to Fall 2008 Artificial Intelligence Midterm Exam INSTRUCTIONS You have 80 minutes. 70 points total. Don t panic! The exam is closed book, closed notes except a one-page crib sheet,

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Fall 2018 Introduction to Artificial Intelligence Practice Final You have approximately 2 hours 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib

More information

Name: UW CSE 473 Final Exam, Fall 2014

Name: UW CSE 473 Final Exam, Fall 2014 P1 P6 Instructions Please answer clearly and succinctly. If an explanation is requested, think carefully before writing. Points may be removed for rambling answers. If a question is unclear or ambiguous,

More information

CS 4100 // artificial intelligence. Recap/midterm review!

CS 4100 // artificial intelligence. Recap/midterm review! CS 4100 // artificial intelligence instructor: byron wallace Recap/midterm review! Attribution: many of these slides are modified versions of those distributed with the UC Berkeley CS188 materials Thanks

More information

Introduction to Fall 2009 Artificial Intelligence Final Exam

Introduction to Fall 2009 Artificial Intelligence Final Exam CS 188 Introduction to Fall 2009 Artificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except a two-page crib sheet. Please use non-programmable calculators

More information

CSL302/612 Artificial Intelligence End-Semester Exam 120 Minutes

CSL302/612 Artificial Intelligence End-Semester Exam 120 Minutes CSL302/612 Artificial Intelligence End-Semester Exam 120 Minutes Name: Roll Number: Please read the following instructions carefully Ø Calculators are allowed. However, laptops or mobile phones are not

More information

Final Exam December 12, 2017

Final Exam December 12, 2017 Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes

More information

CS360 Homework 12 Solution

CS360 Homework 12 Solution CS360 Homework 12 Solution Constraint Satisfaction 1) Consider the following constraint satisfaction problem with variables x, y and z, each with domain {1, 2, 3}, and constraints C 1 and C 2, defined

More information

Introduction to Spring 2006 Artificial Intelligence Practice Final

Introduction to Spring 2006 Artificial Intelligence Practice Final NAME: SID#: Login: Sec: 1 CS 188 Introduction to Spring 2006 Artificial Intelligence Practice Final You have 180 minutes. The exam is open-book, open-notes, no electronics other than basic calculators.

More information

Final Exam December 12, 2017

Final Exam December 12, 2017 Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes

More information

Midterm 2 V1. Introduction to Artificial Intelligence. CS 188 Spring 2015

Midterm 2 V1. Introduction to Artificial Intelligence. CS 188 Spring 2015 S 88 Spring 205 Introduction to rtificial Intelligence Midterm 2 V ˆ You have approximately 2 hours and 50 minutes. ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib

More information

CS221 Practice Midterm #2 Solutions

CS221 Practice Midterm #2 Solutions CS221 Practice Midterm #2 Solutions Summer 2013 Updated 4:00pm, July 24 2 [Deterministic Search] Pacfamily (20 points) Pacman is trying eat all the dots, but he now has the help of his family! There are

More information

CS 188 Fall Introduction to Artificial Intelligence Midterm 1

CS 188 Fall Introduction to Artificial Intelligence Midterm 1 CS 88 Fall 207 Introduction to Artificial Intelligence Midterm You have approximately 0 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark your

More information

Final. CS 188 Fall Introduction to Artificial Intelligence

Final. CS 188 Fall Introduction to Artificial Intelligence CS 188 Fall 2012 Introduction to Artificial Intelligence Final You have approximately 3 hours. The exam is closed book, closed notes except your three one-page crib sheets. Please use non-programmable

More information

ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. S 88 Summer 205 Introduction to rtificial Intelligence Final ˆ You have approximately 2 hours 50 minutes. ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

More information

Mock Exam Künstliche Intelligenz-1. Different problems test different skills and knowledge, so do not get stuck on one problem.

Mock Exam Künstliche Intelligenz-1. Different problems test different skills and knowledge, so do not get stuck on one problem. Name: Matriculation Number: Mock Exam Künstliche Intelligenz-1 January 9., 2017 You have one hour(sharp) for the test; Write the solutions to the sheet. The estimated time for solving this exam is 53 minutes,

More information

CS 188 Fall Introduction to Artificial Intelligence Midterm 2

CS 188 Fall Introduction to Artificial Intelligence Midterm 2 CS 188 Fall 2013 Introduction to rtificial Intelligence Midterm 2 ˆ You have approximately 2 hours and 50 minutes. ˆ The exam is closed book, closed notes except your one-page crib sheet. ˆ Please use

More information

Midterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes.

Midterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes. CS 188 Spring 2013 Introduction to Artificial Intelligence Midterm II ˆ You have approximately 1 hour and 50 minutes. ˆ The exam is closed book, closed notes except a one-page crib sheet. ˆ Please use

More information

CS 7180: Behavioral Modeling and Decisionmaking

CS 7180: Behavioral Modeling and Decisionmaking CS 7180: Behavioral Modeling and Decisionmaking in AI Markov Decision Processes for Complex Decisionmaking Prof. Amy Sliva October 17, 2012 Decisions are nondeterministic In many situations, behavior and

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 88 Fall 208 Introduction to Artificial Intelligence Practice Final You have approximately 2 hours 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

More information

Midterm. Introduction to Artificial Intelligence. CS 188 Summer You have approximately 2 hours and 50 minutes.

Midterm. Introduction to Artificial Intelligence. CS 188 Summer You have approximately 2 hours and 50 minutes. CS 188 Summer 2014 Introduction to Artificial Intelligence Midterm You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your one-page crib sheet. Mark your answers

More information

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes. CS 188 Spring 2013 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except a three-page crib sheet. Please use non-programmable

More information

Final exam of ECE 457 Applied Artificial Intelligence for the Spring term 2007.

Final exam of ECE 457 Applied Artificial Intelligence for the Spring term 2007. Spring 2007 / Page 1 Final exam of ECE 457 Applied Artificial Intelligence for the Spring term 2007. Don t panic. Be sure to write your name and student ID number on every page of the exam. The only materials

More information

1 [15 points] Search Strategies

1 [15 points] Search Strategies Probabilistic Foundations of Artificial Intelligence Final Exam Date: 29 January 2013 Time limit: 120 minutes Number of pages: 12 You can use the back of the pages if you run out of space. strictly forbidden.

More information

Reinforcement Learning and Control

Reinforcement Learning and Control CS9 Lecture notes Andrew Ng Part XIII Reinforcement Learning and Control We now begin our study of reinforcement learning and adaptive control. In supervised learning, we saw algorithms that tried to make

More information

CS 188 Introduction to AI Fall 2005 Stuart Russell Final

CS 188 Introduction to AI Fall 2005 Stuart Russell Final NAME: SID#: Section: 1 CS 188 Introduction to AI all 2005 Stuart Russell inal You have 2 hours and 50 minutes. he exam is open-book, open-notes. 100 points total. Panic not. Mark your answers ON HE EXAM

More information

ARTIFICIAL INTELLIGENCE. Reinforcement learning

ARTIFICIAL INTELLIGENCE. Reinforcement learning INFOB2KI 2018-2019 Utrecht University The Netherlands ARTIFICIAL INTELLIGENCE Reinforcement learning Lecturer: Silja Renooij These slides are part of the INFOB2KI Course Notes available from www.cs.uu.nl/docs/vakken/b2ki/schema.html

More information

ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Summer 2015 Introduction to Artificial Intelligence Midterm 2 ˆ You have approximately 80 minutes. ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

More information

Midterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes.

Midterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes. CS 188 Spring 2013 Introduction to Artificial Intelligence Midterm II ˆ You have approximately 1 hour and 50 minutes. ˆ The exam is closed book, closed notes except a one-page crib sheet. ˆ Please use

More information

Tentamen TDDC17 Artificial Intelligence 20 August 2012 kl

Tentamen TDDC17 Artificial Intelligence 20 August 2012 kl Linköpings Universitet Institutionen för Datavetenskap Patrick Doherty Tentamen TDDC17 Artificial Intelligence 20 August 2012 kl. 08-12 Points: The exam consists of exercises worth 32 points. To pass the

More information

Machine Learning, Midterm Exam

Machine Learning, Midterm Exam 10-601 Machine Learning, Midterm Exam Instructors: Tom Mitchell, Ziv Bar-Joseph Wednesday 12 th December, 2012 There are 9 questions, for a total of 100 points. This exam has 20 pages, make sure you have

More information

Midterm Examination CS540: Introduction to Artificial Intelligence

Midterm Examination CS540: Introduction to Artificial Intelligence Midterm Examination CS540: Introduction to Artificial Intelligence November 1, 2005 Instructor: Jerry Zhu CLOSED BOOK (One letter-size notes allowed. Turn it in with the exam) LAST (FAMILY) NAME: FIRST

More information

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability Due: Thursday 10/15 in 283 Soda Drop Box by 11:59pm (no slip days) Policy: Can be solved in groups (acknowledge collaborators)

More information

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements CS 188: Artificial Intelligence Spring 2011 Lecture 12: Probability 3/2/2011 Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Announcements P3 due on Monday (3/7) at 4:59pm W3 going out

More information

Using first-order logic, formalize the following knowledge:

Using first-order logic, formalize the following knowledge: Probabilistic Artificial Intelligence Final Exam Feb 2, 2016 Time limit: 120 minutes Number of pages: 19 Total points: 100 You can use the back of the pages if you run out of space. Collaboration on the

More information

Today s Outline. Recap: MDPs. Bellman Equations. Q-Value Iteration. Bellman Backup 5/7/2012. CSE 473: Artificial Intelligence Reinforcement Learning

Today s Outline. Recap: MDPs. Bellman Equations. Q-Value Iteration. Bellman Backup 5/7/2012. CSE 473: Artificial Intelligence Reinforcement Learning CSE 473: Artificial Intelligence Reinforcement Learning Dan Weld Today s Outline Reinforcement Learning Q-value iteration Q-learning Exploration / exploitation Linear function approximation Many slides

More information

Final exam of ECE 457 Applied Artificial Intelligence for the Fall term 2007.

Final exam of ECE 457 Applied Artificial Intelligence for the Fall term 2007. Fall 2007 / Page 1 Final exam of ECE 457 Applied Artificial Intelligence for the Fall term 2007. Don t panic. Be sure to write your name and student ID number on every page of the exam. The only materials

More information

Final Exam, Machine Learning, Spring 2009

Final Exam, Machine Learning, Spring 2009 Name: Andrew ID: Final Exam, 10701 Machine Learning, Spring 2009 - The exam is open-book, open-notes, no electronics other than calculators. - The maximum possible score on this exam is 100. You have 3

More information

Be able to define the following terms and answer basic questions about them:

Be able to define the following terms and answer basic questions about them: CS440/ECE448 Section Q Fall 2017 Final Review Be able to define the following terms and answer basic questions about them: Probability o Random variables, axioms of probability o Joint, marginal, conditional

More information

Machine Learning, Fall 2009: Midterm

Machine Learning, Fall 2009: Midterm 10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all

More information

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability Due: Thursday 10/15 in 283 Soda Drop Box by 11:59pm (no slip days) Policy: Can be solved in groups (acknowledge collaborators)

More information

Bayes Nets III: Inference

Bayes Nets III: Inference 1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy

More information

CS 570: Machine Learning Seminar. Fall 2016

CS 570: Machine Learning Seminar. Fall 2016 CS 570: Machine Learning Seminar Fall 2016 Class Information Class web page: http://web.cecs.pdx.edu/~mm/mlseminar2016-2017/fall2016/ Class mailing list: cs570@cs.pdx.edu My office hours: T,Th, 2-3pm or

More information

Final Examination CS 540-2: Introduction to Artificial Intelligence

Final Examination CS 540-2: Introduction to Artificial Intelligence Final Examination CS 540-2: Introduction to Artificial Intelligence May 7, 2017 LAST NAME: SOLUTIONS FIRST NAME: Problem Score Max Score 1 14 2 10 3 6 4 10 5 11 6 9 7 8 9 10 8 12 12 8 Total 100 1 of 11

More information

CSE 546 Final Exam, Autumn 2013

CSE 546 Final Exam, Autumn 2013 CSE 546 Final Exam, Autumn 0. Personal info: Name: Student ID: E-mail address:. There should be 5 numbered pages in this exam (including this cover sheet).. You can use any material you brought: any book,

More information

Introduction to Reinforcement Learning

Introduction to Reinforcement Learning CSCI-699: Advanced Topics in Deep Learning 01/16/2019 Nitin Kamra Spring 2019 Introduction to Reinforcement Learning 1 What is Reinforcement Learning? So far we have seen unsupervised and supervised learning.

More information

Q-Learning in Continuous State Action Spaces

Q-Learning in Continuous State Action Spaces Q-Learning in Continuous State Action Spaces Alex Irpan alexirpan@berkeley.edu December 5, 2015 Contents 1 Introduction 1 2 Background 1 3 Q-Learning 2 4 Q-Learning In Continuous Spaces 4 5 Experimental

More information

Algorithms for Playing and Solving games*

Algorithms for Playing and Solving games* Algorithms for Playing and Solving games* Andrew W. Moore Professor School of Computer Science Carnegie Mellon University www.cs.cmu.edu/~awm awm@cs.cmu.edu 412-268-7599 * Two Player Zero-sum Discrete

More information

CS181 Midterm 2 Practice Solutions

CS181 Midterm 2 Practice Solutions CS181 Midterm 2 Practice Solutions 1. Convergence of -Means Consider Lloyd s algorithm for finding a -Means clustering of N data, i.e., minimizing the distortion measure objective function J({r n } N n=1,

More information

Introduction to Arti Intelligence

Introduction to Arti Intelligence Introduction to Arti Intelligence cial Lecture 4: Constraint satisfaction problems 1 / 48 Constraint satisfaction problems: Today Exploiting the representation of a state to accelerate search. Backtracking.

More information

CS 188: Artificial Intelligence

CS 188: Artificial Intelligence CS 188: Artificial Intelligence Adversarial Search II Instructor: Anca Dragan University of California, Berkeley [These slides adapted from Dan Klein and Pieter Abbeel] Minimax Example 3 12 8 2 4 6 14

More information

Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 25. Minesweeper. Optimal play in Minesweeper

Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 25. Minesweeper. Optimal play in Minesweeper CS 70 Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 25 Minesweeper Our final application of probability is to Minesweeper. We begin by discussing how to play the game optimally; this is probably

More information

CSE 150. Assignment 6 Summer Maximum likelihood estimation. Out: Thu Jul 14 Due: Tue Jul 19

CSE 150. Assignment 6 Summer Maximum likelihood estimation. Out: Thu Jul 14 Due: Tue Jul 19 SE 150. Assignment 6 Summer 2016 Out: Thu Jul 14 ue: Tue Jul 19 6.1 Maximum likelihood estimation A (a) omplete data onsider a complete data set of i.i.d. examples {a t, b t, c t, d t } T t=1 drawn from

More information

Learning in State-Space Reinforcement Learning CIS 32

Learning in State-Space Reinforcement Learning CIS 32 Learning in State-Space Reinforcement Learning CIS 32 Functionalia Syllabus Updated: MIDTERM and REVIEW moved up one day. MIDTERM: Everything through Evolutionary Agents. HW 2 Out - DUE Sunday before the

More information

Learning in Zero-Sum Team Markov Games using Factored Value Functions

Learning in Zero-Sum Team Markov Games using Factored Value Functions Learning in Zero-Sum Team Markov Games using Factored Value Functions Michail G. Lagoudakis Department of Computer Science Duke University Durham, NC 27708 mgl@cs.duke.edu Ronald Parr Department of Computer

More information

CITS4211 Mid-semester test 2011

CITS4211 Mid-semester test 2011 CITS4211 Mid-semester test 2011 Fifty minutes, answer all four questions, total marks 60 Question 1. (12 marks) Briefly describe the principles, operation, and performance issues of iterative deepening.

More information

Decision making, Markov decision processes

Decision making, Markov decision processes Decision making, Markov decision processes Solved tasks Collected by: Jiří Kléma, klema@fel.cvut.cz Spring 2017 The main goal: The text presents solved tasks to support labs in the A4B33ZUI course. 1 Simple

More information

, and rewards and transition matrices as shown below:

, and rewards and transition matrices as shown below: CSE 50a. Assignment 7 Out: Tue Nov Due: Thu Dec Reading: Sutton & Barto, Chapters -. 7. Policy improvement Consider the Markov decision process (MDP) with two states s {0, }, two actions a {0, }, discount

More information

4. What is the probability that the two values differ by 4 or more in absolute value? There are only six

4. What is the probability that the two values differ by 4 or more in absolute value? There are only six 1. Short Questions: 2/2/2/2/2 Provide a clear and concise justification of your answer. In this problem, you roll two balanced six-sided dice. Hint: Draw a picture. 1. What is the probability that the

More information

Distributed Optimization. Song Chong EE, KAIST

Distributed Optimization. Song Chong EE, KAIST Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links

More information

Q1. [24 pts] The OMNIBUS

Q1. [24 pts] The OMNIBUS CS 188 Spring 2011 Introduction to Artificial Intelligence Final Exam Solutions Q1. [24 pts] The OMNIBUS Each question is worth 1 point. Leaving a question blank is worth 0 points. Answering a multiple

More information

Minimax strategies, alpha beta pruning. Lirong Xia

Minimax strategies, alpha beta pruning. Lirong Xia Minimax strategies, alpha beta pruning Lirong Xia Reminder ØProject 1 due tonight Makes sure you DO NOT SEE ERROR: Summation of parsed points does not match ØProject 2 due in two weeks 2 How to find good

More information

Summary. Agenda. Games. Intelligence opponents in game. Expected Value Expected Max Algorithm Minimax Algorithm Alpha-beta Pruning Simultaneous Game

Summary. Agenda. Games. Intelligence opponents in game. Expected Value Expected Max Algorithm Minimax Algorithm Alpha-beta Pruning Simultaneous Game Summary rtificial Intelligence and its applications Lecture 4 Game Playing Search onstraint Satisfaction Problems From start state to goal state onsider constraints Professor Daniel Yeung danyeung@ieee.org

More information

Midterm II. Introduction to Artificial Intelligence. CS 188 Fall You have approximately 3 hours.

Midterm II. Introduction to Artificial Intelligence. CS 188 Fall You have approximately 3 hours. CS 188 Fall 2012 Introduction to Artificial Intelligence Midterm II You have approximately 3 hours. The exam is closed book, closed notes except a one-page crib sheet. Please use non-programmable calculators

More information

Introduction to Machine Learning Midterm, Tues April 8

Introduction to Machine Learning Midterm, Tues April 8 Introduction to Machine Learning 10-701 Midterm, Tues April 8 [1 point] Name: Andrew ID: Instructions: You are allowed a (two-sided) sheet of notes. Exam ends at 2:45pm Take a deep breath and don t spend

More information

Machine Learning, Midterm Exam: Spring 2009 SOLUTION

Machine Learning, Midterm Exam: Spring 2009 SOLUTION 10-601 Machine Learning, Midterm Exam: Spring 2009 SOLUTION March 4, 2009 Please put your name at the top of the table below. If you need more room to work out your answer to a question, use the back of

More information

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012 CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline

More information

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return

More information

Midterm II. Introduction to Artificial Intelligence. CS 188 Fall ˆ You have approximately 3 hours.

Midterm II. Introduction to Artificial Intelligence. CS 188 Fall ˆ You have approximately 3 hours. CS 188 Fall 2012 Introduction to Artificial Intelligence Midterm II ˆ You have approximately 3 hours. ˆ The exam is closed book, closed notes except a one-page crib sheet. ˆ Please use non-programmable

More information

Final Exam. CS 188 Fall Introduction to Artificial Intelligence

Final Exam. CS 188 Fall Introduction to Artificial Intelligence CS 188 Fall 2018 Introduction to rtificial Intelligence Final Exam You have 180 minutes. The time will be projected at the front of the room. You may not leave during the last 10 minutes of the exam. Do

More information

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20.

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20. 10-601 Machine Learning, Midterm Exam: Spring 2008 Please put your name on this cover sheet If you need more room to work out your answer to a question, use the back of the page and clearly mark on the

More information

Reinforcement Learning Wrap-up

Reinforcement Learning Wrap-up Reinforcement Learning Wrap-up Slides courtesy of Dan Klein and Pieter Abbeel University of California, Berkeley [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018 Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections

More information

Final. Introduction to Artificial Intelligence. CS 188 Summer 2014

Final. Introduction to Artificial Intelligence. CS 188 Summer 2014 S 188 Summer 2014 Introduction to rtificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers ON

More information

CS 188 Fall Introduction to Artificial Intelligence Midterm 2

CS 188 Fall Introduction to Artificial Intelligence Midterm 2 CS 188 Fall 2013 Introduction to rtificial Intelligence Midterm 2 ˆ You have approximately 2 hours and 50 minutes. ˆ The exam is closed book, closed notes except your one-page crib sheet. ˆ Please use

More information

This question has three parts, each of which can be answered concisely, but be prepared to explain and justify your concise answer.

This question has three parts, each of which can be answered concisely, but be prepared to explain and justify your concise answer. This question has three parts, each of which can be answered concisely, but be prepared to explain and justify your concise answer. 1. Suppose you have a policy and its action-value function, q, then you

More information

Prioritized Sweeping Converges to the Optimal Value Function

Prioritized Sweeping Converges to the Optimal Value Function Technical Report DCS-TR-631 Prioritized Sweeping Converges to the Optimal Value Function Lihong Li and Michael L. Littman {lihong,mlittman}@cs.rutgers.edu RL 3 Laboratory Department of Computer Science

More information

Partially Observable Markov Decision Processes (POMDPs)

Partially Observable Markov Decision Processes (POMDPs) Partially Observable Markov Decision Processes (POMDPs) Geoff Hollinger Sequential Decision Making in Robotics Spring, 2011 *Some media from Reid Simmons, Trey Smith, Tony Cassandra, Michael Littman, and

More information

TEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY

TEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY TEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY KAREN YE Abstract. In this paper, we differentiate between cold games, which are easier to analyze and play, and hot games, much more difficult in terms

More information

An Algorithm better than AO*?

An Algorithm better than AO*? An Algorithm better than? Blai Bonet Universidad Simón Boĺıvar Caracas, Venezuela Héctor Geffner ICREA and Universitat Pompeu Fabra Barcelona, Spain 7/2005 An Algorithm Better than? B. Bonet and H. Geffner;

More information

FINAL: CS 6375 (Machine Learning) Fall 2014

FINAL: CS 6375 (Machine Learning) Fall 2014 FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for

More information

Examination Artificial Intelligence Module Intelligent Interaction Design December 2014

Examination Artificial Intelligence Module Intelligent Interaction Design December 2014 Examination Artificial Intelligence Module Intelligent Interaction Design December 2014 Introduction This exam is closed book, you may only use a simple calculator (addition, substraction, multiplication

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,

More information

Announcements. CS 188: Artificial Intelligence Fall Adversarial Games. Computing Minimax Values. Evaluation Functions. Recap: Resource Limits

Announcements. CS 188: Artificial Intelligence Fall Adversarial Games. Computing Minimax Values. Evaluation Functions. Recap: Resource Limits CS 188: Artificial Intelligence Fall 2009 Lecture 7: Expectimax Search 9/17/2008 Announcements Written 1: Search and CSPs is up Project 2: Multi-agent Search is up Want a partner? Come to the front after

More information

Final Exam, Fall 2002

Final Exam, Fall 2002 15-781 Final Exam, Fall 22 1. Write your name and your andrew email address below. Name: Andrew ID: 2. There should be 17 pages in this exam (excluding this cover sheet). 3. If you need more room to work

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

CS 188: Artificial Intelligence Fall Announcements

CS 188: Artificial Intelligence Fall Announcements CS 188: Artificial Intelligence Fall 2009 Lecture 7: Expectimax Search 9/17/2008 Dan Klein UC Berkeley Many slides over the course adapted from either Stuart Russell or Andrew Moore 1 Announcements Written

More information

CS 662 Sample Midterm

CS 662 Sample Midterm Name: 1 True/False, plus corrections CS 662 Sample Midterm 35 pts, 5 pts each Each of the following statements is either true or false. If it is true, mark it true. If it is false, correct the statement

More information

Complexity of stochastic branch and bound methods for belief tree search in Bayesian reinforcement learning

Complexity of stochastic branch and bound methods for belief tree search in Bayesian reinforcement learning Complexity of stochastic branch and bound methods for belief tree search in Bayesian reinforcement learning Christos Dimitrakakis Informatics Institute, University of Amsterdam, Amsterdam, The Netherlands

More information

Introduction to Fall 2008 Artificial Intelligence Final Exam

Introduction to Fall 2008 Artificial Intelligence Final Exam CS 188 Introduction to Fall 2008 Artificial Intelligence Final Exam INSTRUCTIONS You have 180 minutes. 100 points total. Don t panic! The exam is closed book, closed notes except a two-page crib sheet,

More information