Introduction to Spring 2009 Artificial Intelligence Midterm Solutions
|
|
- Jonathan Craig
- 5 years ago
- Views:
Transcription
1 S 88 Introduction to Spring 009 rtificial Intelligence Midterm Solutions. (6 points) True/False For the following questions, a correct answer is worth points, no answer is worth point, and an incorrect answer is worth 0 points. ircle true or false to indicate your answer. a) (true or false) If g(s) and h(s) are two admissible heuristics, then their average f(s) = g(s) + h(s) must also be admissible. True. Let h (s) be the true distance from s. We know that g(s) h (s) and h(s) h (s), thus f(s) = g(s) + h(s) h (s) + h (s) = h (s) b) (true or false) For a search problem, the path returned by uniform cost search may change if we add a positive constant to every step cost. True. onsider that there are two paths from the start state (S) to the goal (G), S G and S G. cost(s, ) =, cost(, G) =, and cost(s, G) =. So the optimal path is through. Now, if we add to each of the costs, the optimal path is directly from S to G. Since uniform cost search finds the optimal path, its path will change. c) (true or false) The running-time of an efficient solver for tree-structured constraint satisfaction problems is linear in the number of variables. True. The running time of the algorithm for tree-structured SPs is (n d ), where n is the number of variables and d is the maximum size of any variable s domain. d) (true or false) If h (s) is a consistent heuristic and h (s) is an admissible heuristic, then min(h (s), h (s)) must be consistent. False. For instance, if h (s) be admissible but inconsistent, and h (s) dominate h (s), then min(h (s), h (s)) = h (s), which is inconsistent. e) (true or false) The amount of memory required to run minimax with alpha-beta pruning is (b d ) for branching factor b and depth limit d. True and False (everyone wins). The memory required is only (bd), so we accepted False. However, by definition an algorithm that is (bd) is also (b d ), because denotes upper bounds that may or may not be tight, so technically this statement is True (but not very useful). f) (true or false) In a Markov decision process with discount γ =, the difference in values for two adjacent states is bounded by the reward between them: V (s) V (s ) max a R(s, a, s ). False. Let V (s ) = 0, and R(s, a, s ) = 0 a, but there is an action a which takes s to the terminal state T and gives a reward 00. Thus V (s) 00, but the inequality above says that V (s) 0 0. g) (true or false) Value iteration and policy iteration must always converge to the same policy. True and False (everyone wins). Both algorithms are guaranteed to converge to the optimal policy, so we accepted True. If there are multiple policies that are optimal (meaning they yield the same maximal values for every state), then the algorithms might diverge. Both value iteration and policy iteration will always lead to the same optimal values. h) (true or false) In a Bayes net, if B, then B for some variable other than or B. False. onsider the Bayes net B. learly, B, but B.
2 E! F! G. (5 points) Search: rawler s Escape G G Whilst Pacman was Q-learning, rawler snuck into mediumlassic and stole all the dots. Now, it s trying to escape as quickly as possible. t each time step, G rawler can either move its shoulder G or its elbow one position up or down. Both joints s and e have five total positions each ( through 5) and both begin in position. Upon changing arm positions from (s, e) to (s, e ), rawler s body moves some distance r(s, e, s, e ), where r(s, e, s, e ) meters (negative distance means the crawler moves backwards). rawler must travel 0 meters to reach the exit. N ction: increase elbow position exit r(,,, ) 0 meters In this problem, you will design a search problem for which the optimal solution will allow rawler to escape in as few time steps as possible. (a) ( pt) Define a state space, start state and goal test to represent this problem. state is a -tuple consisting of distance to exit, shoulder position, and elbow position. The start state is (0,, ). Goal test: distance to exit is less than or equal to zero. (b) ( pt) Define the successor function for the start state by listing its (successor state, action, step cost) triples. Use the actions s + and s for increasing and decreasing the shoulder position and e + and e for the elbow. { ((0 r(,,, ),, ), s +, ), ((0 r(,,, ),, ), s, ), ((0 r(,,, ),, ), e +, ), ((0 r(,,, ),, ), e, ) }
3 NME: (c) ( pt) Would depth-first graph search be complete for the search problem you defined? Explain. No, depth-first search would not be complete because the state space is infinite. Depth-first search could forever keep expanding states with increasing distance to the goal. (d) ( pt) Design a non-trivial, admissible heuristic for this problem. Distance to exit divided by is admissible because each step can move rawler at most meters closer to the goal. slightly better heuristic would be to round this heuristic up to the nearest integer. (e) ( pt) rawler s shoulder overheats if it switches direction more than once in any three consecutive time steps, making progress impossible. Describe how you would change your search problem so that an optimal search algorithm would only return solutions that avoid overheating. onsider a reversal to be s + followed by s or s followed by s +. To avoid these, we can extend the state space to include the last two actions. We then remove from the output of the successor function any triple that includes the action sequences (s +, s, s + ) or (s, s +, s ). We also accepted solutions that gave a very high step cost to these successors.
4 . (0 points) SPs: onstraining Graph Structure Hired as a consultant, you built a Bayes net model of heeseboard Pizza customers. Last night, a corporate spy from Zachary s Pizza deleted the directions from! all your arcs.!! E You still have the undirected n! graph of your! E! E n-! model and a list of dependencies. You now have to recover the direction of the arrows to build a graph that allows for these dependencies. Note: Y means is not independent of Y. E! B Distribution properties (dependencies): G!!! n! E! E n-! B F! E! G F! G B F B G a) ( pt) Given the first constraint only, circle all the topologies that are allowed for the triple (,, G). G G G G P! T! P! G G r(s 0, e 0, s 0, e 0 +) G G b) ( pt) Formulate direction-recovery for this graph as a SP with explicit binary constraints only. The D variables and values are provided for you. The variable E stands for the direction of the arc between N ction: increase and the center, where out means the arrow points outward elbow position (toward ), and in means the arrow points T! inward (toward ). Variables: E, E B, E F, E G r(,,, ) Values: out, in 0 meters onstraints: n explicit constraint N lists all legal tuples of values are listed for a tuple of variables. (E, E G ) {(in, out), (out, out), (out, in)} (E B, E F ) {(in, out), (out, out), (out, in)} (E B, E G ) {(in, in)} exit 0 meters c) ( pt) Draw the constraint graph for this SP. E E G E B E F d) ( pt) fter selecting E = in, cross out all values eliminated from E B, E F and E G by forward checking. E E B E F E G in out in out in out in Forward checking removes the values of variables that directly contradict the assignment E = in. e) ( pt) ross out all values eliminated by arc consistency applied before any backtracking search. E E B E F E G out in out in out in out in We first remove E B = out because it is incompatible with any value for E G. Likewise, we remove E G = out. Now, we can remove E = in based on E G and E F = in based on E B. f) ( pt) Solve this SP, then add the correct directions to the arcs of the graph at the top of the page. E = out, E B = in, E F = out, and E G = in. exit
5 NME: 5 Now consider a chain structured graph over variables,..., n. Edges E,..., E n connect adjacent variables, but their directions are again unknown. B D!!! E n!! E! E n-! g) ( pt) Using only binary constraints and two-valued variables, formulate a SP that is satisfied by only and all networks that can represent a distribution where n. Describe what your variables mean in terms of direction of the arrows in the network. Variables: Variables E,..., E n correspond to the directions of these edges and take values {left, right}. onstraints: (E i, E i+ ) (right, left), for i {0,..., n }. h) (6 pt) Using only unary, binary andternary ( variable) constraints andtwo-valued variables, formulate a SP that is satisfied by only and all networks that enforce n. Describe what your variables mean in terms of direction of the arrows in the network. Hint: You will have to introduce additional variables. Variables: Variables E,..., E n correspond to the directions of these edges and take values {left, right}. I,..., I n can take values {T, F }.,..., n can take values {T, F }. onstraints: Intuitively, we want I i to be T if i i+ i+, i.e. the triple i, i+, i+ is inactive. lso, we want i to be T if either I i or i is T, i.e. if the current triple is inactive or if any of the prior triples was inactive. Finally, we want n to be T which would imply that there is some inactive triple. (I i, E i, E i+ ) {(T, right, left), (F, right, right), (F, left, right), (F, left, left)} for i {,..., n }. (I, ) {(T, T ), (F, F )} ( i, I i, i ) {(F, F, F ), (F, T, T ), (T, F, T ), (T, T, T )} for i {,..., n } n = true
6 6. (0 points) Multi-agent Search: onnect- In the game of onnect-, players and alternate moves, dropping their symbols into columns,,, or. Three-in-a-row wins the game, horizontally, vertically or diagonally. plays first. (a) ( pt) What is the maximum branching factor of minimax search for onnect-?. (b) ( pt) What is the maximum tree depth in plies? 6. ccording to our terminology, a search ply in a multi-agent search problem includes one move for every agent. Because the textbook differs in its definition, we also accepted. (c) ( pt) Give a reasonably tight upper bound on the number of terminal states. Several answers were accepted: is a reasonable bound on the number of leaves of the search tree. is a reasonable bound on the number of total states in the game, because each square can be empty, contain an or contain an. tighter bound is 5, which follows from the observation that there are only 5 ways to legally fill a column from the bottom up with s and s. Bounds based on the assumption that the entire board would be filled with s and s were not accepted, even if they in fact overestimated the true number of terminal states. (d) ( pt) Draw the game tree starting from the board shown above right (with to play next). You may abbreviate states by drawing only the upper-right region circled. The root node is drawn for you.!"#!"# $#!"# $#!%#!"#!%# (e) ( pt) is the maximizer, while is the minimizer. s utility for terminal states is k when wins and k when wins, where k is for a horizontal -in-a-row win (as in above left), for a vertical win, and for a diagonal win. tie has value 0. Label each node of your tree with its minimax value. (f) ( pt) ircle all nodes of your tree that will not be explored when using alpha-beta pruning and a move ordering that maximizes the number of nodes pruned.
7 NME: 7 5. ( points) MDPs: Robot Soccer soccer robot is on a fast break toward the goal, starting in position. From positions through, it can either shoot (S) or dribble the ball forward (D); from it can only shoot. If it shoots, it either scores a goal (state G) or misses (state M). If it dribbles, it either advances a square or loses the ball, ending up in M. D Goal In this MDP, the states are,,,, G and M, where G and M are terminal states. The transition model depends on the parameter y, which is the probability of dribbling success. ssume a discount of γ =. T (k, S, G) = k 6 T (k, S, M) = k 6 for k {,,, } T (k, D, k + ) = y T (k, D, M) = y for k {,, } R(k, S, G) = for k {,,, }, and rewards are 0 for all other transitions (a) ( pt) What is V π () for the policy π that always shoots? V π () = T (, S, G)R(, S, G) + T (, S, M)R(, S, M) = 6 (b) ( pt) What is Q (, D) in terms of y? Q (, D) = T (, D, )(R(, D, ) + V ()) + T (, D, M)R(, D, M) = T (, D, )V () = T (, D, )Q (, S) = T (, D, )(T (, S, G)R(, S, G) + T (, S, M)R(, S, m)) = T (, D, )T (, S, G)R(, S, G) = y (c) ( pt) Using y =, complete the first two iterations of value iteration. i Vi () V i () V i () V i () (d) ( pt) fter how many iterations will value iteration compute the optimal values for all states? fter iterations, the values will have converged when y =. bove, only V () has not yet converged. We note that for y >, a fourth iteration would be required because a fast break has up to four transitions. (e) ( pt) For what range of values of y is Q (, S) Q (, D)? Q (, S) Q (, D) T (, S, G) T (, D, ) T (, S, G) y y 0
8 8 The dribble success probability y in fact depend on the presence or absence of a defending robot, D. has no way of detecting whether D is present, but does know some statistical properties of its environment. D is present of the time. When D is absent, y =. When D is present, y =. (f) ( pt) What is the posterior probability that D is present, given that Dribbles twice successfully from to, then Shoots from state and scores. We can use Bayes rule, where D is a random variable denoting the presence of D, and e is the evidence that dribbled twice and scored. P (d e) = P (e d) P (d) P (e) P (e) = P (e d) P (d) + P (e d) P ( d) P (e d) = P (e d) = P (e) = + 9 = 96 P (d e) = 96 / 96 = (g) ( pt) What transition model should use in order to correctly compute its maximum expected reward when it doesn t know whether or not D is present? To maximize expected total reward, the agent should model the situation as accurately as possible. We accepted two answers for T (k, D, k + ). ne answer is to claim that the agent should use its marginal belief that each dribble will be successful, summing over the two possibilities of D s presence or absense. In this case, T (k, D, k + ) = + = 5 7, and T (k, D, M) =. nother acceptable answer would be to update the beliefs of the agent after every successful dribble. Hence, while T (, D, ) = 5 as above, reaching state implies a successful previous dribble, and so another successful dribble is more likely. In particular, let, and indicate successful first, second and third dribbles, while D is the presence of the defender. Since D, we have P (x x ) = d P (d x )P (x d). omputing similarly to part (f), P (d x ) = 5, so T (, D, ) = = 0. We can compute T (, D, ) similarly. is conditionally independent of both and given D, so P (x x, x ) = d P (d x, x )P (x d). Using our result from (f), P (d x, x ) =, and so T (, D, ) = + 9 = 9. Since a dribble either succeeds or fails, T (k, D, M) = T (k, D, k + ) for all k. Shooting probabilities are unchanged by the presence of the defender. T (k, S, G) = k 6 T (k, S, M) = k 6 for k {,,, } T (k, D, k + ) = 5 T (k, D, M) = 7 for k {,, } R T (k, S, G) = k 6 T (k, S, M) = k 6 for k {,,, } T (, D, ) = 5 T (, D, ) = 0 T (, D, ) = 9 T (, D, M) = 7 T (, D, M) = 9 0 T (, D, M) = 5
9 NME: 9 (h) ( pt) What is the optimal policy π when doesn t know whether or not D is present? Using T (k, D, k + ) = 5, we find that π = { : S, : S, : S, : S}. We showed from part (e) that for any y <, shooting is preferable to dribbling from state. Therefore, we know that V () = Q (, S). We can perform similar computations for states and : V () = Q (, S) = π () = S Q (, S) = Q (, D) = 5 V () = 5 V () = max(q (, S), Q (, D)) = π () = S Q (, S) = 6 Q (, D) = 5 V () = 5 6 V () = max(q (, S), Q (, D)) = 6 π () = S Under the second answer for part (g), similar computations give π = { : S, : S, : S, : S}. V () = Q (, S) = Q (, D) = 9 V () = V () = max(q (, S), Q (, D)) = π () = S Q (, S) = Q (, D) = 0 V () = V () = max(q (, S), Q (, D)) = π () = S Q (, S) = 6 Q (, D) = 5 V () = 5 6 V () = max(q (, S), Q (, D)) = 6 π () = S
10 B 0 D 6. (8 points) Bayes Nets: The Mind of Pacman Pacman doesn t just eat dots. He also enjoys pears, apples, carrots, nuts, eggs and toast. Each morning, he chooses to eat some subset of these foods. His preferences are modeled by a Bayes net with this structure. E P! T! N a) ( pt) Factor the probability that Pacman chooses to eat only apples and nuts in terms of conditional probabilities from this Bayes net. P ( p, a, c, n, e, t) = P (a) P ( p a) P ( c a) P (n p, c) P ( t c) P ( e n, t) b) ( pt) For each of the following properties, circle whether they are true, false or unknown of the distribution P (P,,, N, E, T ) for this Bayes net. Without the conditional probability tables, we do not know if any two variables are dependent. From the network topology, we can only conclude true or unknown for the questions below. N T (true, false, unknown) unknown: T N is an active path. P E N (true, false, unknown) unknown: P T E is an active path. P N, (true, false, unknown) unknown: P and N are directly connected. E, N (true, false, unknown) true: ll three paths from E to are inactive. c) ( pt) If the arrow from T to E were reversed, list two conditional independence properties that would be true under the new graph that were not guaranteed under the original graph. The triple T E is no longer active when T is unknown, therefore P E N and E N and E N The triple N E T is no longer active when E is known, therefore T N, E and T P, E and T, E There are further independence assumptions based on these 6 that condition on more variables, such as E N, P. d) ( pt) You discover that P (a c, t) > P (a c). Suggest how you might change the network in response. This inequality implies that T by definition, but in the network above, and T are conditionally independent given. So, the network must be changed to allow for this dependence. dding an arc from to T would fix the problem. Note: adding an arc from T to creates a cycle, which is not allowed in a Bayes net. ther acceptable solutions exist, like reversing the direction of the arc from T to.
11 NME: e) ( pt) You now want to compute the distribution P () using variable elimination. List the factors that remain before and after eliminating the variable N. Before: The initial factors are just the conditional probability tables of the Bayes net. P (), P (P ), P ( ), P (T ), P (N P, ), P (E T, N) fter: First, all factors that include the variable N are joined together, yielding P (E, N P,, T ) Next, the variable N is summed out of this new factor, yielding P (E P,, T ) The remaining factors include this new factor and the unused original factors. P (), P (P ), P ( ), P (T ), P (E P,, T ) Referring to the final factor as m(e, P,, T ) (like in the textbook) was also accepted. f) ( pt) Pacman s new diet allows only fruit (P and ) to be eaten, but Pacman only follows the diet occasionally. dd the new variable D (for whether he follows the diet) to the network below by adding arcs. Briefly justify your answer. P! P! D D T! N T! N The best answer is the one on the left, which allows us to express changes to all food preferences based on the diet. ther answers were accepted with appropriate justification, such as the network on the right, which specifies that the diet specifically disallows, N, E and T. onnecting D only to P and/or is problematic: for instance, observing P and would make D independent of, but the diet should still affect whether carrots () are allowed even when and P are known. g) ( pt) Given the Bayes net and the factors below, fill in the table for P (D a) or state that there is not enough information. From the tables P (D) and P ( D), the entries of the factor P (D = a) can be computed from Bayes rule (which always holds): no additional information about the properties of the distributions are required. The trick is to observe that while P ( a) is not explicitly given, it can be computed from P (D) and P ( D) using the product rule: a,d : P (a, d) = P (d)p (a d). omputing P ( a) in this way is equivalent to applying the normalization trick. D P (D) d 0. d 0.6 D P ( D) d a 0.8 d a 0. d a 0.5 d a 0.5 D P (D = a) P ( a d) P (d) d a P ( a d) P ( d)+p ( a d) P (d) = = 9 P ( a d) P ( d) d a P ( a d) P ( d)+p ( a d) P (d) = = 5 9
Introduction to Spring 2009 Artificial Intelligence Midterm Exam
S 188 Introduction to Spring 009 rtificial Intelligence Midterm Exam INSTRUTINS You have 3 hours. The exam is closed book, closed notes except a one-page crib sheet. Please use non-programmable calculators
More informationCS221 Practice Midterm
CS221 Practice Midterm Autumn 2012 1 ther Midterms The following pages are excerpts from similar classes midterms. The content is similar to what we ve been covering this quarter, so that it should be
More informationFinal. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.
CS 188 Spring 2014 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Fall 2015 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Spring 2017 Introduction to Artificial Intelligence Midterm V2 You have approximately 80 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark
More informationCS 188 Introduction to Fall 2007 Artificial Intelligence Midterm
NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.
More informationSection Handout 5 Solutions
CS 188 Spring 2019 Section Handout 5 Solutions Approximate Q-Learning With feature vectors, we can treat values of states and q-states as linear value functions: V (s) = w 1 f 1 (s) + w 2 f 2 (s) +...
More informationFinal. CS 188 Fall Introduction to Artificial Intelligence
S 188 Fall 2012 Introduction to rtificial Intelligence Final You have approximately 3 hours. The exam is closed book, closed notes except your three one-page crib sheets. Please use non-programmable calculators
More informationIntroduction to Fall 2011 Artificial Intelligence Final Exam
CS 188 Introduction to Fall 2011 rtificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except two pages of crib sheets. Please use non-programmable calculators
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Spring 2017 Introduction to Artificial Intelligence Midterm V2 You have approximately 80 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark
More informationIntroduction to Fall 2008 Artificial Intelligence Midterm Exam
CS 188 Introduction to Fall 2008 Artificial Intelligence Midterm Exam INSTRUCTIONS You have 80 minutes. 70 points total. Don t panic! The exam is closed book, closed notes except a one-page crib sheet,
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Fall 2018 Introduction to Artificial Intelligence Practice Final You have approximately 2 hours 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib
More informationName: UW CSE 473 Final Exam, Fall 2014
P1 P6 Instructions Please answer clearly and succinctly. If an explanation is requested, think carefully before writing. Points may be removed for rambling answers. If a question is unclear or ambiguous,
More informationCS 4100 // artificial intelligence. Recap/midterm review!
CS 4100 // artificial intelligence instructor: byron wallace Recap/midterm review! Attribution: many of these slides are modified versions of those distributed with the UC Berkeley CS188 materials Thanks
More informationIntroduction to Fall 2009 Artificial Intelligence Final Exam
CS 188 Introduction to Fall 2009 Artificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except a two-page crib sheet. Please use non-programmable calculators
More informationCSL302/612 Artificial Intelligence End-Semester Exam 120 Minutes
CSL302/612 Artificial Intelligence End-Semester Exam 120 Minutes Name: Roll Number: Please read the following instructions carefully Ø Calculators are allowed. However, laptops or mobile phones are not
More informationFinal Exam December 12, 2017
Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes
More informationCS360 Homework 12 Solution
CS360 Homework 12 Solution Constraint Satisfaction 1) Consider the following constraint satisfaction problem with variables x, y and z, each with domain {1, 2, 3}, and constraints C 1 and C 2, defined
More informationIntroduction to Spring 2006 Artificial Intelligence Practice Final
NAME: SID#: Login: Sec: 1 CS 188 Introduction to Spring 2006 Artificial Intelligence Practice Final You have 180 minutes. The exam is open-book, open-notes, no electronics other than basic calculators.
More informationFinal Exam December 12, 2017
Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes
More informationMidterm 2 V1. Introduction to Artificial Intelligence. CS 188 Spring 2015
S 88 Spring 205 Introduction to rtificial Intelligence Midterm 2 V ˆ You have approximately 2 hours and 50 minutes. ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib
More informationCS221 Practice Midterm #2 Solutions
CS221 Practice Midterm #2 Solutions Summer 2013 Updated 4:00pm, July 24 2 [Deterministic Search] Pacfamily (20 points) Pacman is trying eat all the dots, but he now has the help of his family! There are
More informationCS 188 Fall Introduction to Artificial Intelligence Midterm 1
CS 88 Fall 207 Introduction to Artificial Intelligence Midterm You have approximately 0 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark your
More informationFinal. CS 188 Fall Introduction to Artificial Intelligence
CS 188 Fall 2012 Introduction to Artificial Intelligence Final You have approximately 3 hours. The exam is closed book, closed notes except your three one-page crib sheets. Please use non-programmable
More informationˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
S 88 Summer 205 Introduction to rtificial Intelligence Final ˆ You have approximately 2 hours 50 minutes. ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
More informationMock Exam Künstliche Intelligenz-1. Different problems test different skills and knowledge, so do not get stuck on one problem.
Name: Matriculation Number: Mock Exam Künstliche Intelligenz-1 January 9., 2017 You have one hour(sharp) for the test; Write the solutions to the sheet. The estimated time for solving this exam is 53 minutes,
More informationCS 188 Fall Introduction to Artificial Intelligence Midterm 2
CS 188 Fall 2013 Introduction to rtificial Intelligence Midterm 2 ˆ You have approximately 2 hours and 50 minutes. ˆ The exam is closed book, closed notes except your one-page crib sheet. ˆ Please use
More informationMidterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes.
CS 188 Spring 2013 Introduction to Artificial Intelligence Midterm II ˆ You have approximately 1 hour and 50 minutes. ˆ The exam is closed book, closed notes except a one-page crib sheet. ˆ Please use
More informationCS 7180: Behavioral Modeling and Decisionmaking
CS 7180: Behavioral Modeling and Decisionmaking in AI Markov Decision Processes for Complex Decisionmaking Prof. Amy Sliva October 17, 2012 Decisions are nondeterministic In many situations, behavior and
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 88 Fall 208 Introduction to Artificial Intelligence Practice Final You have approximately 2 hours 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
More informationMidterm. Introduction to Artificial Intelligence. CS 188 Summer You have approximately 2 hours and 50 minutes.
CS 188 Summer 2014 Introduction to Artificial Intelligence Midterm You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your one-page crib sheet. Mark your answers
More informationFinal. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.
CS 188 Spring 2013 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except a three-page crib sheet. Please use non-programmable
More informationFinal exam of ECE 457 Applied Artificial Intelligence for the Spring term 2007.
Spring 2007 / Page 1 Final exam of ECE 457 Applied Artificial Intelligence for the Spring term 2007. Don t panic. Be sure to write your name and student ID number on every page of the exam. The only materials
More information1 [15 points] Search Strategies
Probabilistic Foundations of Artificial Intelligence Final Exam Date: 29 January 2013 Time limit: 120 minutes Number of pages: 12 You can use the back of the pages if you run out of space. strictly forbidden.
More informationReinforcement Learning and Control
CS9 Lecture notes Andrew Ng Part XIII Reinforcement Learning and Control We now begin our study of reinforcement learning and adaptive control. In supervised learning, we saw algorithms that tried to make
More informationCS 188 Introduction to AI Fall 2005 Stuart Russell Final
NAME: SID#: Section: 1 CS 188 Introduction to AI all 2005 Stuart Russell inal You have 2 hours and 50 minutes. he exam is open-book, open-notes. 100 points total. Panic not. Mark your answers ON HE EXAM
More informationARTIFICIAL INTELLIGENCE. Reinforcement learning
INFOB2KI 2018-2019 Utrecht University The Netherlands ARTIFICIAL INTELLIGENCE Reinforcement learning Lecturer: Silja Renooij These slides are part of the INFOB2KI Course Notes available from www.cs.uu.nl/docs/vakken/b2ki/schema.html
More informationˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Summer 2015 Introduction to Artificial Intelligence Midterm 2 ˆ You have approximately 80 minutes. ˆ The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
More informationMidterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes.
CS 188 Spring 2013 Introduction to Artificial Intelligence Midterm II ˆ You have approximately 1 hour and 50 minutes. ˆ The exam is closed book, closed notes except a one-page crib sheet. ˆ Please use
More informationTentamen TDDC17 Artificial Intelligence 20 August 2012 kl
Linköpings Universitet Institutionen för Datavetenskap Patrick Doherty Tentamen TDDC17 Artificial Intelligence 20 August 2012 kl. 08-12 Points: The exam consists of exercises worth 32 points. To pass the
More informationMachine Learning, Midterm Exam
10-601 Machine Learning, Midterm Exam Instructors: Tom Mitchell, Ziv Bar-Joseph Wednesday 12 th December, 2012 There are 9 questions, for a total of 100 points. This exam has 20 pages, make sure you have
More informationMidterm Examination CS540: Introduction to Artificial Intelligence
Midterm Examination CS540: Introduction to Artificial Intelligence November 1, 2005 Instructor: Jerry Zhu CLOSED BOOK (One letter-size notes allowed. Turn it in with the exam) LAST (FAMILY) NAME: FIRST
More informationCS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability
CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability Due: Thursday 10/15 in 283 Soda Drop Box by 11:59pm (no slip days) Policy: Can be solved in groups (acknowledge collaborators)
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 12: Probability 3/2/2011 Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Announcements P3 due on Monday (3/7) at 4:59pm W3 going out
More informationUsing first-order logic, formalize the following knowledge:
Probabilistic Artificial Intelligence Final Exam Feb 2, 2016 Time limit: 120 minutes Number of pages: 19 Total points: 100 You can use the back of the pages if you run out of space. Collaboration on the
More informationToday s Outline. Recap: MDPs. Bellman Equations. Q-Value Iteration. Bellman Backup 5/7/2012. CSE 473: Artificial Intelligence Reinforcement Learning
CSE 473: Artificial Intelligence Reinforcement Learning Dan Weld Today s Outline Reinforcement Learning Q-value iteration Q-learning Exploration / exploitation Linear function approximation Many slides
More informationFinal exam of ECE 457 Applied Artificial Intelligence for the Fall term 2007.
Fall 2007 / Page 1 Final exam of ECE 457 Applied Artificial Intelligence for the Fall term 2007. Don t panic. Be sure to write your name and student ID number on every page of the exam. The only materials
More informationFinal Exam, Machine Learning, Spring 2009
Name: Andrew ID: Final Exam, 10701 Machine Learning, Spring 2009 - The exam is open-book, open-notes, no electronics other than calculators. - The maximum possible score on this exam is 100. You have 3
More informationBe able to define the following terms and answer basic questions about them:
CS440/ECE448 Section Q Fall 2017 Final Review Be able to define the following terms and answer basic questions about them: Probability o Random variables, axioms of probability o Joint, marginal, conditional
More informationMachine Learning, Fall 2009: Midterm
10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all
More informationCS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability
CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability Due: Thursday 10/15 in 283 Soda Drop Box by 11:59pm (no slip days) Policy: Can be solved in groups (acknowledge collaborators)
More informationBayes Nets III: Inference
1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy
More informationCS 570: Machine Learning Seminar. Fall 2016
CS 570: Machine Learning Seminar Fall 2016 Class Information Class web page: http://web.cecs.pdx.edu/~mm/mlseminar2016-2017/fall2016/ Class mailing list: cs570@cs.pdx.edu My office hours: T,Th, 2-3pm or
More informationFinal Examination CS 540-2: Introduction to Artificial Intelligence
Final Examination CS 540-2: Introduction to Artificial Intelligence May 7, 2017 LAST NAME: SOLUTIONS FIRST NAME: Problem Score Max Score 1 14 2 10 3 6 4 10 5 11 6 9 7 8 9 10 8 12 12 8 Total 100 1 of 11
More informationCSE 546 Final Exam, Autumn 2013
CSE 546 Final Exam, Autumn 0. Personal info: Name: Student ID: E-mail address:. There should be 5 numbered pages in this exam (including this cover sheet).. You can use any material you brought: any book,
More informationIntroduction to Reinforcement Learning
CSCI-699: Advanced Topics in Deep Learning 01/16/2019 Nitin Kamra Spring 2019 Introduction to Reinforcement Learning 1 What is Reinforcement Learning? So far we have seen unsupervised and supervised learning.
More informationQ-Learning in Continuous State Action Spaces
Q-Learning in Continuous State Action Spaces Alex Irpan alexirpan@berkeley.edu December 5, 2015 Contents 1 Introduction 1 2 Background 1 3 Q-Learning 2 4 Q-Learning In Continuous Spaces 4 5 Experimental
More informationAlgorithms for Playing and Solving games*
Algorithms for Playing and Solving games* Andrew W. Moore Professor School of Computer Science Carnegie Mellon University www.cs.cmu.edu/~awm awm@cs.cmu.edu 412-268-7599 * Two Player Zero-sum Discrete
More informationCS181 Midterm 2 Practice Solutions
CS181 Midterm 2 Practice Solutions 1. Convergence of -Means Consider Lloyd s algorithm for finding a -Means clustering of N data, i.e., minimizing the distortion measure objective function J({r n } N n=1,
More informationIntroduction to Arti Intelligence
Introduction to Arti Intelligence cial Lecture 4: Constraint satisfaction problems 1 / 48 Constraint satisfaction problems: Today Exploiting the representation of a state to accelerate search. Backtracking.
More informationCS 188: Artificial Intelligence
CS 188: Artificial Intelligence Adversarial Search II Instructor: Anca Dragan University of California, Berkeley [These slides adapted from Dan Klein and Pieter Abbeel] Minimax Example 3 12 8 2 4 6 14
More informationDiscrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 25. Minesweeper. Optimal play in Minesweeper
CS 70 Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 25 Minesweeper Our final application of probability is to Minesweeper. We begin by discussing how to play the game optimally; this is probably
More informationCSE 150. Assignment 6 Summer Maximum likelihood estimation. Out: Thu Jul 14 Due: Tue Jul 19
SE 150. Assignment 6 Summer 2016 Out: Thu Jul 14 ue: Tue Jul 19 6.1 Maximum likelihood estimation A (a) omplete data onsider a complete data set of i.i.d. examples {a t, b t, c t, d t } T t=1 drawn from
More informationLearning in State-Space Reinforcement Learning CIS 32
Learning in State-Space Reinforcement Learning CIS 32 Functionalia Syllabus Updated: MIDTERM and REVIEW moved up one day. MIDTERM: Everything through Evolutionary Agents. HW 2 Out - DUE Sunday before the
More informationLearning in Zero-Sum Team Markov Games using Factored Value Functions
Learning in Zero-Sum Team Markov Games using Factored Value Functions Michail G. Lagoudakis Department of Computer Science Duke University Durham, NC 27708 mgl@cs.duke.edu Ronald Parr Department of Computer
More informationCITS4211 Mid-semester test 2011
CITS4211 Mid-semester test 2011 Fifty minutes, answer all four questions, total marks 60 Question 1. (12 marks) Briefly describe the principles, operation, and performance issues of iterative deepening.
More informationDecision making, Markov decision processes
Decision making, Markov decision processes Solved tasks Collected by: Jiří Kléma, klema@fel.cvut.cz Spring 2017 The main goal: The text presents solved tasks to support labs in the A4B33ZUI course. 1 Simple
More information, and rewards and transition matrices as shown below:
CSE 50a. Assignment 7 Out: Tue Nov Due: Thu Dec Reading: Sutton & Barto, Chapters -. 7. Policy improvement Consider the Markov decision process (MDP) with two states s {0, }, two actions a {0, }, discount
More information4. What is the probability that the two values differ by 4 or more in absolute value? There are only six
1. Short Questions: 2/2/2/2/2 Provide a clear and concise justification of your answer. In this problem, you roll two balanced six-sided dice. Hint: Draw a picture. 1. What is the probability that the
More informationDistributed Optimization. Song Chong EE, KAIST
Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links
More informationQ1. [24 pts] The OMNIBUS
CS 188 Spring 2011 Introduction to Artificial Intelligence Final Exam Solutions Q1. [24 pts] The OMNIBUS Each question is worth 1 point. Leaving a question blank is worth 0 points. Answering a multiple
More informationMinimax strategies, alpha beta pruning. Lirong Xia
Minimax strategies, alpha beta pruning Lirong Xia Reminder ØProject 1 due tonight Makes sure you DO NOT SEE ERROR: Summation of parsed points does not match ØProject 2 due in two weeks 2 How to find good
More informationSummary. Agenda. Games. Intelligence opponents in game. Expected Value Expected Max Algorithm Minimax Algorithm Alpha-beta Pruning Simultaneous Game
Summary rtificial Intelligence and its applications Lecture 4 Game Playing Search onstraint Satisfaction Problems From start state to goal state onsider constraints Professor Daniel Yeung danyeung@ieee.org
More informationMidterm II. Introduction to Artificial Intelligence. CS 188 Fall You have approximately 3 hours.
CS 188 Fall 2012 Introduction to Artificial Intelligence Midterm II You have approximately 3 hours. The exam is closed book, closed notes except a one-page crib sheet. Please use non-programmable calculators
More informationIntroduction to Machine Learning Midterm, Tues April 8
Introduction to Machine Learning 10-701 Midterm, Tues April 8 [1 point] Name: Andrew ID: Instructions: You are allowed a (two-sided) sheet of notes. Exam ends at 2:45pm Take a deep breath and don t spend
More informationMachine Learning, Midterm Exam: Spring 2009 SOLUTION
10-601 Machine Learning, Midterm Exam: Spring 2009 SOLUTION March 4, 2009 Please put your name at the top of the table below. If you need more room to work out your answer to a question, use the back of
More informationOutline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012
CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline
More informationAnnouncements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic
CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return
More informationMidterm II. Introduction to Artificial Intelligence. CS 188 Fall ˆ You have approximately 3 hours.
CS 188 Fall 2012 Introduction to Artificial Intelligence Midterm II ˆ You have approximately 3 hours. ˆ The exam is closed book, closed notes except a one-page crib sheet. ˆ Please use non-programmable
More informationFinal Exam. CS 188 Fall Introduction to Artificial Intelligence
CS 188 Fall 2018 Introduction to rtificial Intelligence Final Exam You have 180 minutes. The time will be projected at the front of the room. You may not leave during the last 10 minutes of the exam. Do
More informationMachine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20.
10-601 Machine Learning, Midterm Exam: Spring 2008 Please put your name on this cover sheet If you need more room to work out your answer to a question, use the back of the page and clearly mark on the
More informationReinforcement Learning Wrap-up
Reinforcement Learning Wrap-up Slides courtesy of Dan Klein and Pieter Abbeel University of California, Berkeley [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationSection Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018
Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections
More informationFinal. Introduction to Artificial Intelligence. CS 188 Summer 2014
S 188 Summer 2014 Introduction to rtificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers ON
More informationCS 188 Fall Introduction to Artificial Intelligence Midterm 2
CS 188 Fall 2013 Introduction to rtificial Intelligence Midterm 2 ˆ You have approximately 2 hours and 50 minutes. ˆ The exam is closed book, closed notes except your one-page crib sheet. ˆ Please use
More informationThis question has three parts, each of which can be answered concisely, but be prepared to explain and justify your concise answer.
This question has three parts, each of which can be answered concisely, but be prepared to explain and justify your concise answer. 1. Suppose you have a policy and its action-value function, q, then you
More informationPrioritized Sweeping Converges to the Optimal Value Function
Technical Report DCS-TR-631 Prioritized Sweeping Converges to the Optimal Value Function Lihong Li and Michael L. Littman {lihong,mlittman}@cs.rutgers.edu RL 3 Laboratory Department of Computer Science
More informationPartially Observable Markov Decision Processes (POMDPs)
Partially Observable Markov Decision Processes (POMDPs) Geoff Hollinger Sequential Decision Making in Robotics Spring, 2011 *Some media from Reid Simmons, Trey Smith, Tony Cassandra, Michael Littman, and
More informationTEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY
TEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY KAREN YE Abstract. In this paper, we differentiate between cold games, which are easier to analyze and play, and hot games, much more difficult in terms
More informationAn Algorithm better than AO*?
An Algorithm better than? Blai Bonet Universidad Simón Boĺıvar Caracas, Venezuela Héctor Geffner ICREA and Universitat Pompeu Fabra Barcelona, Spain 7/2005 An Algorithm Better than? B. Bonet and H. Geffner;
More informationFINAL: CS 6375 (Machine Learning) Fall 2014
FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for
More informationExamination Artificial Intelligence Module Intelligent Interaction Design December 2014
Examination Artificial Intelligence Module Intelligent Interaction Design December 2014 Introduction This exam is closed book, you may only use a simple calculator (addition, substraction, multiplication
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,
More informationAnnouncements. CS 188: Artificial Intelligence Fall Adversarial Games. Computing Minimax Values. Evaluation Functions. Recap: Resource Limits
CS 188: Artificial Intelligence Fall 2009 Lecture 7: Expectimax Search 9/17/2008 Announcements Written 1: Search and CSPs is up Project 2: Multi-agent Search is up Want a partner? Come to the front after
More informationFinal Exam, Fall 2002
15-781 Final Exam, Fall 22 1. Write your name and your andrew email address below. Name: Andrew ID: 2. There should be 17 pages in this exam (excluding this cover sheet). 3. If you need more room to work
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationCS 188: Artificial Intelligence Fall Announcements
CS 188: Artificial Intelligence Fall 2009 Lecture 7: Expectimax Search 9/17/2008 Dan Klein UC Berkeley Many slides over the course adapted from either Stuart Russell or Andrew Moore 1 Announcements Written
More informationCS 662 Sample Midterm
Name: 1 True/False, plus corrections CS 662 Sample Midterm 35 pts, 5 pts each Each of the following statements is either true or false. If it is true, mark it true. If it is false, correct the statement
More informationComplexity of stochastic branch and bound methods for belief tree search in Bayesian reinforcement learning
Complexity of stochastic branch and bound methods for belief tree search in Bayesian reinforcement learning Christos Dimitrakakis Informatics Institute, University of Amsterdam, Amsterdam, The Netherlands
More informationIntroduction to Fall 2008 Artificial Intelligence Final Exam
CS 188 Introduction to Fall 2008 Artificial Intelligence Final Exam INSTRUCTIONS You have 180 minutes. 100 points total. Don t panic! The exam is closed book, closed notes except a two-page crib sheet,
More information