Final Exam. 1 True or False (15 Points)
|
|
- Abraham Williamson
- 5 years ago
- Views:
Transcription
1 Final Exam Submit by Oct. 16, :59pm EST Please submit early, and update your submission if you want to make changes. Do not wait to the last minute to submit: we reserve the right not to accept late submissions. Please look over the exam before starting, and ask questions during the class period; we will answer questions as quickly as we can during class, but later in the day it may take substantially longer. For the true or false questions: you may print out this page, or answer on a separate page. If you do the latter, please make sure you label your answers clearly for which question they belong to. 1 True or False (15 Points) 1. (circle one: T F ) For any finite-dimensional vector space, the size of its basis is equal to its dimension. 2. (circle one: T F ) For any finite-dimensional vector space, there is always a unique basis. 3. (circle one: T F ) The perceptron algorithm always converges to a binary classifier that has zero error rate. 4. (circle one: T F ) {} and {{}} are the same set. 5. (circle one: T F ) {} is a subset of {{}}. 6. (circle one: T F ) {} is an element of {{}}. 7. (circle one: T F ) Every vector space contains a zero vector. 8. (circle one: T F ) A vector space can have more than one zero vector. 9. (circle one: T F ) A system of n linear equations with n unknowns has at least one solution. 10. (circle one: T F ) A system of n linear equations with n unknowns has at most one solution. 11. (circle one: T F ) Matrices can be used to represent any translation or rotation on the vector space R n. 12. (circle one: T F ) Any combination of rotations on the vector space R n can be combined into a single matrix. 13. (circle one: T F ) Every square matrix is invertible. 14. (circle one: T F ) Every symmetric matrix is invertible. 15. (circle one: T F ) Matrix multiplication is associative and commutative. 1
2 2 Linear algebra (15 Points) 1. Consider the vector space of all 2 by 2 matrices with real entries, denoted R 2 2. In this space, we define vector addition to be componentwise matrix addition, and we define scalar multiplication to be componentwise scalar multiplication. What is the dimension of this vector space? 2. In the same space, consider the following vectors: [ ] [ ] [ ] b 1 =, b =, b = 0 1 Are they sufficient to form a basis for R 2 2? If not, which vector(s) should be added? 3. Now consider the space of symmetric 2 2 matrices, with the same addition and scalar multiplication operations. Write down a basis for this space. 3 Transformations (15 Points) Write the matrices that correspond to the following transformations: 1. M : R 2 R 3 defined by M(x, y) = (2x + 3y, x y, 9x) 2. M : P 5 P 5 defined by (Mf)(t) = f (t), where P n is the space of degree-n polynomials with the standard basis (1, t, t 2,..., t n ) 4 Matrix Inverses (20 Points) In this question, all matrices are square. Be sure to justify every step in your proofs: indicate which previous statements the step follows from, and what rule(s) you re using to make this step. 1. We know that if A is an invertible matrix, then there exists a matrix A 1 such that AA 1 = I and A 1 A = I. Assume we know that there exist matrices B and C, such that BA = I and AC = I. Prove that B = C. (Hint: try beginning with the product BAC.) 2. If matrices X and Y are invertible, prove that (XY ) 1 = Y 1 X 1. 5 Kernel Perceptron (20 Points) As discussed in class, we can generalize the perceptron algorithm to work on any Hilbert space (complete inner product space) the result is called a kernel perceptron. In this question we ll work out how to implement the kernel percepton when our feature vector φ(x) is in R q with q large (so that it s impractical to work with vectors like φ(x) explicitly). To do so, we will assume that we have a subroutine to 2
3 compute φ(x) φ(x ) for any vectors x, x efficiently (i.e., without ever building a vector in R q explicitly); this operation is called a kernel function. Assume we have observed input vectors x i R p with labels y i { 1, 1} for i = 1... n. Also assume we have a feature function φ : R p R q. Then the ordinary perceptron algorithm works as follows: we start with weight vector w = 0 and bias b = 0. We run through the data repeatedly. As long as our predictions are correct, we leave w and b alone. If we make a mistake, we update (if the true label is positive, y i = 1), or w w + φ(x i ) b b + 1 w w φ(x i ) b b 1 (if the true label is negative, y i = 1). If on some iteration w φ(x i ) + b = 0, we consider this example to be a mistake, no matter what the correct sign y i is. Because of the form of the perceptron update, we can see that the weights are always a linear combination of the labeled training data, i.e., w = n a j y j φ(x j ). (1) j=1 The only way we ever use the weight vector w is to make predictions: ŷ i = sign(w φ(x i ) + b) (2) ( n ) = sign a j y j φ(x j ) φ(x i ) + b (3) j=1 So, we never need to store w itself, but only the coefficient vector a = (a 1, a 2,...). The predictions depend on the data only through dot products of the form K ij = φ(x j ) φ(x i ) which we assumed above that we can compute efficiently. K R n n is called the kernel matrix or Gram matrix; for our purposes, we can compute it once before starting the algorithm. The kernel perceptron is this version of the algorithm: the one that stores a and b instead of w and b. For this problem, we will use one-dimensional observations x R and a quadratic feature function φ(x) = (x, x 2 ). (Yes, we know that q = 2 is not actually too large to work with.) You are given four data points with observed x values of 1, 0.1, 0.1, 1, and labels (y values) of 1, 1, 1, Give an expression for the value of φ(x i ) φ(x j ) in terms of x i and x j. 2. Compute the kernel matrix K for this data. 3
4 3. Starting with parameters a = (0, 0, 0, 0) and b = 0, simulate two full passes of the kernel perceptron through the data (in the order given), reporting the prediction for each data point and the value of a and b after processing each data point. For a trivial update (one which doesn t change the parameters), please just write same instead of repeating the values of a and b. 4. What is the error rate of the final parameters a and b on this data set? That is, what fraction of examples are classified incorrectly? 5. What is the lowest error rate that we could achieve on this data with an ordinary linear perceptron ŷ = sign(wx + b)? Hint: for part 3, you should only need to make nontrivial updates on approximately half of the steps. 6 Differentials (15 Points) Consider constant vectors a, b, c, d, e and a variable matrix X. Suppose y = a ln(exp(b T Xc) + exp(d T Xe)). Derive an expression for the differential dy in terms of dx. Show your work. If you computed the ordinary derivative dy, what would be the number of modes dx of the resulting tensor? A Extra credit: Probability (10 Points) In this problem you will work out a probability distribution for statistics about demographics and likes for traveling. First, denote X as a random variable indicating gender and Y as a random variable indicating a like for traveling. Suppose we have: Male Female P (X) = 3/5 2/5 Like Traveling Dislike Traveling P (Y ) = 2/3 1/3 P (Z X): Male Female Age /6 1/12 Age /6 1/4 Age /6 2/3 1. Construct the joint probability table P (X, Y ) assuming that these two random variables are independent. 4
5 2. Now suppose you are given additional statistics: a conditional distribution of Z, age, given X, gender. With this extra information (along with the assumptions in part 1): Compute the joint table P (X, Z). What is the probability of a person being female given that the age is 40 50? What is the probability that the age is 20 30? 5
HOMEWORK 4: SVMS AND KERNELS
HOMEWORK 4: SVMS AND KERNELS CMU 060: MACHINE LEARNING (FALL 206) OUT: Sep. 26, 206 DUE: 5:30 pm, Oct. 05, 206 TAs: Simon Shaolei Du, Tianshu Ren, Hsiao-Yu Fish Tung Instructions Homework Submission: Submit
More informationFinal Exam. Problem Score Problem Score 1 /10 8 /10 2 /10 9 /10 3 /10 10 /10 4 /10 11 /10 5 /10 12 /10 6 /10 13 /10 7 /10 Total /130
EE103/CME103: Introduction to Matrix Methods December 9 2015 S. Boyd Final Exam You may not use any books, notes, or computer programs (e.g., Julia). Throughout this exam we use standard mathematical notation;
More informationTest 3, Linear Algebra
Test 3, Linear Algebra Dr. Adam Graham-Squire, Fall 2017 Name: I pledge that I have neither given nor received any unauthorized assistance on this exam. (signature) DIRECTIONS 1. Don t panic. 2. Show all
More informationCSE 151 Machine Learning. Instructor: Kamalika Chaudhuri
CSE 151 Machine Learning Instructor: Kamalika Chaudhuri Linear Classification Given labeled data: (xi, feature vector yi) label i=1,..,n where y is 1 or 1, find a hyperplane to separate from Linear Classification
More informationThe value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.
Math 410 Homework Problems In the following pages you will find all of the homework problems for the semester. Homework should be written out neatly and stapled and turned in at the beginning of class
More informationMATH 300, Second Exam REVIEW SOLUTIONS. NOTE: You may use a calculator for this exam- You only need something that will perform basic arithmetic.
MATH 300, Second Exam REVIEW SOLUTIONS NOTE: You may use a calculator for this exam- You only need something that will perform basic arithmetic. [ ] [ ] 2 2. Let u = and v =, Let S be the parallelegram
More informationExam 3 MATH Calculus I
Trinity College December 03, 2015 MATH 131-01 Calculus I By signing below, you attest that you have neither given nor received help of any kind on this exam. Signature: Printed Name: Instructions: Show
More informationThis is a closed book exam. No notes or calculators are permitted. We will drop your lowest scoring question for you.
Math 54 Fall 2017 Practice Exam 1 Exam date: 9/26/17 Time Limit: 80 Minutes Name: Student ID: GSI or Section: This exam contains 6 pages (including this cover page) and 7 problems. Problems are printed
More informationName (NetID): (1 Point)
CS446: Machine Learning (D) Spring 2017 March 16 th, 2017 This is a closed book exam. Everything you need in order to solve the problems is supplied in the body of this exam. This exam booklet contains
More informationDEPARTMENT OF MATHEMATICS
Name(Last/First): GUID: DEPARTMENT OF MATHEMATICS Ma322005(Sathaye) - Final Exam Spring 2017 May 3, 2017 DO NOT TURN THIS PAGE UNTIL YOU ARE INSTRUCTED TO DO SO. Be sure to show all work and justify your
More informationDEPARTMENT OF MATHEMATICS
Name(Last/First): GUID: DEPARTMENT OF MATHEMATICS Ma322005(Sathaye) - Final Exam Spring 2017 May 3, 2017 DO NOT TURN THIS PAGE UNTIL YOU ARE INSTRUCTED TO DO SO. Be sure to show all work and justify your
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationEach new feature uses a pair of the original features. Problem: Mapping usually leads to the number of features blow up!
Feature Mapping Consider the following mapping φ for an example x = {x 1,...,x D } φ : x {x1,x 2 2,...,x 2 D,,x 2 1 x 2,x 1 x 2,...,x 1 x D,...,x D 1 x D } It s an example of a quadratic mapping Each new
More informationNo books, notes, any calculator, or electronic devices are allowed on this exam. Show all of your steps in each answer to receive a full credit.
MTH 309-001 Fall 2016 Exam 1 10/05/16 Name (Print): PID: READ CAREFULLY THE FOLLOWING INSTRUCTION Do not open your exam until told to do so. This exam contains 7 pages (including this cover page) and 7
More informationFINAL EXAM: FALL 2013 CS 6375 INSTRUCTOR: VIBHAV GOGATE
FINAL EXAM: FALL 2013 CS 6375 INSTRUCTOR: VIBHAV GOGATE You are allowed a two-page cheat sheet. You are also allowed to use a calculator. Answer the questions in the spaces provided on the question sheets.
More informationMachine Learning (CSE 446): Multi-Class Classification; Kernel Methods
Machine Learning (CSE 446): Multi-Class Classification; Kernel Methods Sham M Kakade c 2018 University of Washington cse446-staff@cs.washington.edu 1 / 12 Announcements HW3 due date as posted. make sure
More informationAssignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.
Assignment 1 Math 5341 Linear Algebra Review Give complete answers to each of the following questions Show all of your work Note: You might struggle with some of these questions, either because it has
More informationThis is a closed book exam. No notes or calculators are permitted. We will drop your lowest scoring question for you.
Math 54 Fall 2017 Practice Final Exam Exam date: 12/14/17 Time Limit: 170 Minutes Name: Student ID: GSI or Section: This exam contains 9 pages (including this cover page) and 10 problems. Problems are
More informationChapter 2. Matrix Arithmetic. Chapter 2
Matrix Arithmetic Matrix Addition and Subtraction Addition and subtraction act element-wise on matrices. In order for the addition/subtraction (A B) to be possible, the two matrices A and B must have the
More informationName: Final Exam MATH 3320
Name: Final Exam MATH 3320 Directions: Make sure to show all necessary work to receive full credit. If you need extra space please use the back of the sheet with appropriate labeling. (1) State the following
More informationAdvanced Introduction to Machine Learning
10-715 Advanced Introduction to Machine Learning Homework Due Oct 15, 10.30 am Rules Please follow these guidelines. Failure to do so, will result in loss of credit. 1. Homework is due on the due date
More informationLinear Algebra. Min Yan
Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2010 Lecture 24: Perceptrons and More! 4/22/2010 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements W7 due tonight [this is your last written for
More information5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.
Linear Algebra - Test File - Spring Test # For problems - consider the following system of equations. x + y - z = x + y + 4z = x + y + 6z =.) Solve the system without using your calculator..) Find the
More informationSignature. Printed Name. Math 312 Hour Exam 1 Jerry L. Kazdan March 5, :00 1:20
Signature Printed Name Math 312 Hour Exam 1 Jerry L. Kazdan March 5, 1998 12:00 1:20 Directions: This exam has three parts. Part A has 4 True-False questions, Part B has 3 short answer questions, and Part
More informationThis is a closed book exam. No notes or calculators are permitted. We will drop your lowest scoring question for you.
Math 54 Fall 2017 Practice Exam 2 Exam date: 10/31/17 Time Limit: 80 Minutes Name: Student ID: GSI or Section: This exam contains 7 pages (including this cover page) and 7 problems. Problems are printed
More informationProblem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30
Problem Set 2 MAS 622J/1.126J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 30 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain
More information2. (10 pts) How many vectors are in the null space of the matrix A = 0 1 1? (i). Zero. (iv). Three. (ii). One. (v).
Exam 3 MAS 3105 Applied Linear Algebra, Spring 2018 (Clearly!) Print Name: Apr 10, 2018 Read all of what follows carefully before starting! 1. This test has 7 problems and is worth 110 points. Please be
More informationMachine Learning, Fall 2011: Homework 5
0-60 Machine Learning, Fall 0: Homework 5 Machine Learning Department Carnegie Mellon University Due:??? Instructions There are 3 questions on this assignment. Please submit your completed homework to
More information2.3. VECTOR SPACES 25
2.3. VECTOR SPACES 25 2.3 Vector Spaces MATH 294 FALL 982 PRELIM # 3a 2.3. Let C[, ] denote the space of continuous functions defined on the interval [,] (i.e. f(x) is a member of C[, ] if f(x) is continuous
More informationKernel Methods. Charles Elkan October 17, 2007
Kernel Methods Charles Elkan elkan@cs.ucsd.edu October 17, 2007 Remember the xor example of a classification problem that is not linearly separable. If we map every example into a new representation, then
More informationFamily Feud Review. Linear Algebra. October 22, 2013
Review Linear Algebra October 22, 2013 Question 1 Let A and B be matrices. If AB is a 4 7 matrix, then determine the dimensions of A and B if A has 19 columns. Answer 1 Answer A is a 4 19 matrix, while
More informationCOMS F18 Homework 3 (due October 29, 2018)
COMS 477-2 F8 Homework 3 (due October 29, 208) Instructions Submit your write-up on Gradescope as a neatly typeset (not scanned nor handwritten) PDF document by :59 PM of the due date. On Gradescope, be
More informationSummer Session Practice Final Exam
Math 2F Summer Session 25 Practice Final Exam Time Limit: Hours Name (Print): Teaching Assistant This exam contains pages (including this cover page) and 9 problems. Check to see if any pages are missing.
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table
More informationEssential Question: What is a complex number, and how can you add, subtract, and multiply complex numbers? Explore Exploring Operations Involving
Locker LESSON 3. Complex Numbers Name Class Date 3. Complex Numbers Common Core Math Standards The student is expected to: N-CN. Use the relation i = 1 and the commutative, associative, and distributive
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationDEPARTMENT OF MATHEMATICS
Name(Last/First): GUID: DEPARTMENT OF MATHEMATICS Ma322005(Sathaye) - Final Exam Spring 2017 May 3, 2017 DO NOT TURN THIS PAGE UNTIL YOU ARE INSTRUCTED TO DO SO. Be sure to show all work and justify your
More informationwhich are not all zero. The proof in the case where some vector other than combination of the other vectors in S is similar.
It follows that S is linearly dependent since the equation is satisfied by which are not all zero. The proof in the case where some vector other than combination of the other vectors in S is similar. is
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationPhysics 331 Introduction to Numerical Techniques in Physics
Physics 331 Introduction to Numerical Techniques in Physics Instructor: Joaquín Drut Lecture 12 Last time: Polynomial interpolation: basics; Lagrange interpolation. Today: Quick review. Formal properties.
More informationCalculating determinants for larger matrices
Day 26 Calculating determinants for larger matrices We now proceed to define det A for n n matrices A As before, we are looking for a function of A that satisfies the product formula det(ab) = det A det
More informationCSE 546 Final Exam, Autumn 2013
CSE 546 Final Exam, Autumn 0. Personal info: Name: Student ID: E-mail address:. There should be 5 numbered pages in this exam (including this cover sheet).. You can use any material you brought: any book,
More informationMath 113 Final Exam: Solutions
Math 113 Final Exam: Solutions Thursday, June 11, 2013, 3.30-6.30pm. 1. (25 points total) Let P 2 (R) denote the real vector space of polynomials of degree 2. Consider the following inner product on P
More informationCS446: Machine Learning Fall Final Exam. December 6 th, 2016
CS446: Machine Learning Fall 2016 Final Exam December 6 th, 2016 This is a closed book exam. Everything you need in order to solve the problems is supplied in the body of this exam. This exam booklet contains
More informationMath 110, Spring 2015: Midterm Solutions
Math 11, Spring 215: Midterm Solutions These are not intended as model answers ; in many cases far more explanation is provided than would be necessary to receive full credit. The goal here is to make
More informationCOMS 4771 Introduction to Machine Learning. Nakul Verma
COMS 4771 Introduction to Machine Learning Nakul Verma Announcements HW1 due next lecture Project details are available decide on the group and topic by Thursday Last time Generative vs. Discriminative
More informationMath Computation Test 1 September 26 th, 2016 Debate: Computation vs. Theory Whatever wins, it ll be Huuuge!
Math 5- Computation Test September 6 th, 6 Debate: Computation vs. Theory Whatever wins, it ll be Huuuge! Name: Answer Key: Making Math Great Again Be sure to show your work!. (8 points) Consider the following
More informationReduction to the associated homogeneous system via a particular solution
June PURDUE UNIVERSITY Study Guide for the Credit Exam in (MA 5) Linear Algebra This study guide describes briefly the course materials to be covered in MA 5. In order to be qualified for the credit, one
More informationMidterm: CS 6375 Spring 2015 Solutions
Midterm: CS 6375 Spring 2015 Solutions The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for an
More informationJim Lambers MAT 610 Summer Session Lecture 1 Notes
Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra
More informationBasic Linear Algebra in MATLAB
Basic Linear Algebra in MATLAB 9.29 Optional Lecture 2 In the last optional lecture we learned the the basic type in MATLAB is a matrix of double precision floating point numbers. You learned a number
More informationChoose three of: Choose three of: Choose three of:
MATH Final Exam (Version ) Solutions July 8, 8 S. F. Ellermeyer Name Instructions. Remember to include all important details of your work. You will not get full credit (or perhaps even any partial credit)
More informationis Use at most six elementary row operations. (Partial
MATH 235 SPRING 2 EXAM SOLUTIONS () (6 points) a) Show that the reduced row echelon form of the augmented matrix of the system x + + 2x 4 + x 5 = 3 x x 3 + x 4 + x 5 = 2 2x + 2x 3 2x 4 x 5 = 3 is. Use
More informationReview Packet 1 B 11 B 12 B 13 B = B 21 B 22 B 23 B 31 B 32 B 33 B 41 B 42 B 43
Review Packet. For each of the following, write the vector or matrix that is specified: a. e 3 R 4 b. D = diag{, 3, } c. e R 3 d. I. For each of the following matrices and vectors, give their dimension.
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More informationProblem 1A. Find the volume of the solid given by x 2 + z 2 1, y 2 + z 2 1. (Hint: 1. Solution: The volume is 1. Problem 2A.
Problem 1A Find the volume of the solid given by x 2 + z 2 1, y 2 + z 2 1 (Hint: 1 1 (something)dz) Solution: The volume is 1 1 4xydz where x = y = 1 z 2 This integral has value 16/3 Problem 2A Let f(x)
More informationa b = a T b = a i b i (1) i=1 (Geometric definition) The dot product of two Euclidean vectors a and b is defined by a b = a b cos(θ a,b ) (2)
This is my preperation notes for teaching in sections during the winter 2018 quarter for course CSE 446. Useful for myself to review the concepts as well. More Linear Algebra Definition 1.1 (Dot Product).
More informationMachine Learning Lecture 6 Note
Machine Learning Lecture 6 Note Compiled by Abhi Ashutosh, Daniel Chen, and Yijun Xiao February 16, 2016 1 Pegasos Algorithm The Pegasos Algorithm looks very similar to the Perceptron Algorithm. In fact,
More informationExam 2 MAS 3105 Applied Linear Algebra, Spring 2018
Exam 2 MAS 3105 Applied Linear Algebra, Spring 2018 (Clearly!) Print Name: Mar 8, 2018 Read all of what follows carefully before starting! 1. This test has 6 problems and is worth 100 points. Please be
More informationMAT Linear Algebra Collection of sample exams
MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More informationMachine Learning Practice Page 2 of 2 10/28/13
Machine Learning 10-701 Practice Page 2 of 2 10/28/13 1. True or False Please give an explanation for your answer, this is worth 1 pt/question. (a) (2 points) No classifier can do better than a naive Bayes
More information1 Machine Learning Concepts (16 points)
CSCI 567 Fall 2018 Midterm Exam DO NOT OPEN EXAM UNTIL INSTRUCTED TO DO SO PLEASE TURN OFF ALL CELL PHONES Problem 1 2 3 4 5 6 Total Max 16 10 16 42 24 12 120 Points Please read the following instructions
More informationMath 224, Fall 2007 Exam 3 Thursday, December 6, 2007
Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007 You have 1 hour and 20 minutes. No notes, books, or other references. You are permitted to use Maple during this exam, but you must start with a blank
More informationCS446: Machine Learning Spring Problem Set 4
CS446: Machine Learning Spring 2017 Problem Set 4 Handed Out: February 27 th, 2017 Due: March 11 th, 2017 Feel free to talk to other members of the class in doing the homework. I am more concerned that
More informationLinear Algebra (Review) Volker Tresp 2017
Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.
More informationCIS 520: Machine Learning Oct 09, Kernel Methods
CIS 520: Machine Learning Oct 09, 207 Kernel Methods Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture They may or may not cover all the material discussed
More informationLinear Regression (continued)
Linear Regression (continued) Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 6, 2017 1 / 39 Outline 1 Administration 2 Review of last lecture 3 Linear regression
More informationSupport Vector Machines
Two SVM tutorials linked in class website (please, read both): High-level presentation with applications (Hearst 1998) Detailed tutorial (Burges 1998) Support Vector Machines Machine Learning 10701/15781
More informationTHEODORE VORONOV DIFFERENTIABLE MANIFOLDS. Fall Last updated: November 26, (Under construction.)
4 Vector fields Last updated: November 26, 2009. (Under construction.) 4.1 Tangent vectors as derivations After we have introduced topological notions, we can come back to analysis on manifolds. Let M
More informationDesigning Information Devices and Systems I Spring 2019 Lecture Notes Note 2
EECS 6A Designing Information Devices and Systems I Spring 9 Lecture Notes Note Vectors and Matrices In the previous note, we introduced vectors and matrices as a way of writing systems of linear equations
More informationEXAM 2 REVIEW DAVID SEAL
EXAM 2 REVIEW DAVID SEAL 3. Linear Systems and Matrices 3.2. Matrices and Gaussian Elimination. At this point in the course, you all have had plenty of practice with Gaussian Elimination. Be able to row
More informationCMU CS 462/662 (INTRO TO COMPUTER GRAPHICS) HOMEWORK 0.0 MATH REVIEW/PREVIEW LINEAR ALGEBRA
CMU CS 462/662 (INTRO TO COMPUTER GRAPHICS) HOMEWORK 0.0 MATH REVIEW/PREVIEW LINEAR ALGEBRA Andrew ID: ljelenak August 25, 2018 This assignment reviews basic mathematical tools you will use throughout
More informationMATH10212 Linear Algebra B Homework Week 4
MATH22 Linear Algebra B Homework Week 4 Students are strongly advised to acquire a copy of the Textbook: D. C. Lay Linear Algebra and its Applications. Pearson, 26. ISBN -52-2873-4. Normally, homework
More informationProblem Point Value Points
Math 70 TUFTS UNIVERSITY October 12, 2015 Linear Algebra Department of Mathematics Sections 1 and 2 Exam I Instructions: No notes or books are allowed. All calculators, cell phones, or other electronic
More informationAnswers in blue. If you have questions or spot an error, let me know. 1. Find all matrices that commute with A =. 4 3
Answers in blue. If you have questions or spot an error, let me know. 3 4. Find all matrices that commute with A =. 4 3 a b If we set B = and set AB = BA, we see that 3a + 4b = 3a 4c, 4a + 3b = 3b 4d,
More informationMachine Learning, Fall 2009: Midterm
10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all
More informationI. Multiple Choice Questions (Answer any eight)
Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam - Course Instructor : Prashanth L.A. Date : Sep-24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY
More informationFinal Exam Practice 3, May 8, 2018 Math 21b, Spring Name:
Final Exam Practice 3, May 8, 8 Math b, Spring 8 Name: MWF 9 Oliver Knill MWF Jeremy Hahn MWF Hunter Spink MWF Matt Demers MWF Yu-Wen Hsu MWF Ben Knudsen MWF Sander Kupers MWF Hakim Walker TTH Ana Balibanu
More informationMath 3013 Problem Set 4
(e) W = {x, 3x, 4x 3, 5x 4 x i R} in R 4 Math 33 Problem Set 4 Problems from.6 (pgs. 99- of text):,3,5,7,9,,7,9,,35,37,38. (Problems,3,4,7,9 in text). Determine whether the indicated subset is a subspace
More informationMath 313 (Linear Algebra) Exam 2 - Practice Exam
Name: Student ID: Section: Instructor: Math 313 (Linear Algebra) Exam 2 - Practice Exam Instructions: For questions which require a written answer, show all your work. Full credit will be given only if
More informationNo books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question.
Math 304 Final Exam (May 8) Spring 206 No books, no notes, no calculators. You must show work, unless the question is a true/false, yes/no, or fill-in-the-blank question. Name: Section: Question Points
More informationAlgebra Exam. Solutions and Grading Guide
Algebra Exam Solutions and Grading Guide You should use this grading guide to carefully grade your own exam, trying to be as objective as possible about what score the TAs would give your responses. Full
More informationMark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.
CS 189 Spring 2015 Introduction to Machine Learning Midterm You have 80 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. No calculators or electronic items.
More informationLinear Algebra (MATH ) Spring 2011 Final Exam Practice Problem Solutions
Linear Algebra (MATH 4) Spring 2 Final Exam Practice Problem Solutions Instructions: Try the following on your own, then use the book and notes where you need help. Afterwards, check your solutions with
More informationLinear Algebra (Review) Volker Tresp 2018
Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c
More informationHomework 6: Image Completion using Mixture of Bernoullis
Homework 6: Image Completion using Mixture of Bernoullis Deadline: Wednesday, Nov. 21, at 11:59pm Submission: You must submit two files through MarkUs 1 : 1. a PDF file containing your writeup, titled
More informationMATH 260 LINEAR ALGEBRA EXAM II Fall 2013 Instructions: The use of built-in functions of your calculator, such as det( ) or RREF, is prohibited.
MAH 60 LINEAR ALGEBRA EXAM II Fall 0 Instructions: he use of built-in functions of your calculator, such as det( ) or RREF, is prohibited ) For the matrix find: a) M and C b) M 4 and C 4 ) Evaluate the
More informationSupervised Learning Coursework
Supervised Learning Coursework John Shawe-Taylor Tom Diethe Dorota Glowacka November 30, 2009; submission date: noon December 18, 2009 Abstract Using a series of synthetic examples, in this exercise session
More informationUNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013
UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 Exam policy: This exam allows two one-page, two-sided cheat sheets; No other materials. Time: 2 hours. Be sure to write your name and
More information5 Group theory. 5.1 Binary operations
5 Group theory This section is an introduction to abstract algebra. This is a very useful and important subject for those of you who will continue to study pure mathematics. 5.1 Binary operations 5.1.1
More informationJanuary 2016 Qualifying Examination
January 2016 Qualifying Examination If you have any difficulty with the wording of the following problems please contact the supervisor immediately. All persons responsible for these problems, in principle,
More informationMATH2210 Notebook 3 Spring 2018
MATH2210 Notebook 3 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 3 MATH2210 Notebook 3 3 3.1 Vector Spaces and Subspaces.................................
More informationStatistics 253/317 Introduction to Probability Models. Winter Midterm Exam Friday, Feb 8, 2013
Statistics 253/317 Introduction to Probability Models Winter 2014 - Midterm Exam Friday, Feb 8, 2013 Student Name (print): (a) Do not sit directly next to another student. (b) This is a closed-book, closed-note
More informationMath 2210Q (Roby) Practice Midterm #1 Solutions Fall 2017
Math Q (Roby) Practice Midterm # Solutions Fall 7 SHOW ALL YOUR WORK! Make sure you give reasons to support your answers. If you have any questions, do not hesitate to ask! For this exam no calculators
More informationPhysics 2080 Extra Credit Due March 15, 2011
Physics 2080 Extra Credit Due March 15, 2011 This assignment covers chapters 15-19 and is worth 10 points extra credit on the final exam. Print these pages and do all work here. Complete the questions
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationQ1 Q2 Q3 Q4 Tot Letr Xtra
Mathematics 54.1 Final Exam, 12 May 2011 180 minutes, 90 points NAME: ID: GSI: INSTRUCTIONS: You must justify your answers, except when told otherwise. All the work for a question should be on the respective
More informationMath 416, Spring 2010 Matrix multiplication; subspaces February 2, 2010 MATRIX MULTIPLICATION; SUBSPACES. 1. Announcements
Math 416, Spring 010 Matrix multiplication; subspaces February, 010 MATRIX MULTIPLICATION; SUBSPACES 1 Announcements Office hours on Wednesday are cancelled because Andy will be out of town If you email
More informationFinal Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2
Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch
More information