Matrix Multiplication
|
|
- Joshua Townsend
- 5 years ago
- Views:
Transcription
1 Matrix Multiplication Example (Cost Analysis, 45 in 2.4) The Mundo Candy Company makes three types of chocolate candy: Cheery Cherry, Mucho Mocha, and Almond Delight. The company produces its products in San Diego, Mexico City, and Managua using two main ingredients: chocolate and sugar. a. Each kilogram of Cheery Cherry requires 0.5kg of sugar and 0.2kg of chocolate; each kilogram of Mucho Mocha requires 0.4kg of sugar and 0.3kg of chocolate; and each kilogram of Almond Delight requires 0.3kg of sugar and 0.3kg of chocolate. Put this information into a 2 3 matrix called A, labeling the rows and columns. b. The cost of 1kg of sugar is $4 in San Diego, $2 in Mexico City, and $1 in Managua. The cost of 1kg of chocolate is $3 in San Diego, $5 in Mexico City, and $7 in Managua. Put this information into a matrix called B. c. Write a matrix C which represents the ingredient cost of producing each type of candy in each city. Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 1 / 13
2 EXERCISE Exercise How much does it cost to produce CC in Mexico City? What about the other cities? COST = (c of S) (amt of S) + (c of C) (amt of C) CC in MC: How do we organize the calculation so we get all the answers at once? Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 2 / 13
3 ANSWER Multiplication of matrices: cs cc SD 4 3 MC 2 5 M 1 7 CC MM AD as ac CC MM AD SD MC M Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 3 / 13
4 Transition Matrix, Illustration EXPLAINING TRANSITION MATRICES USING THE WEATHER Suppose in a certain region on any given day, the chance of the weather changing depends only on what the weather was doing on the previous day. To keep things simple, we will say that the weather has two states - either wet or dry. Let s say that the probability of being dry on any day is conditional ONLY upon whether or not it was dry the day before. Specifically lets assume that the probability of being dry on any day, given it was dry on the previous n, is 0.3. Using standard conditional probability notation we can write that Pr(D t D y ) = 0.3. Since dry is the complement of wet, we can say that Pr(W t D y ) = 0.7. Further, suppose we assume that the probability of being wet today, given it was wet yesterday is 0.6. Thus we can similarly write that Pr(W t W y ) = 0.6 and Pr(D t W y ) = 0.4. Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 4 / 13
5 Questions Given that it is WET today, what is the probability that it will be wet two days from now? The probability that it will be wet today is 0.2. What is the probability it will be wet two days from now? What is the transition matrix for 2, 3,..., n days from now? Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 5 / 13
6 State Diagram The situation is illustrated in what is known as a state diagram. Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 6 / 13
7 Tree Diagram We can also describe this with a tree diagram. Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 7 / 13
8 Transition Matrix, Formal Definition Definition A square n n matrix p p 1n... p 1n... p nn is called a Transition Matrix if the 0 p ij 1 and the sums along rows add up to 1. An 1 n matrix, also called an n row vector [ v1... v n ] is called a probability vector if 0 v i 1, and the sum of the coordinates is 1. Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 8 / 13
9 P(1 1) P(2 1)... P(n 1) M =.... P(1 n) P(2 1)... P(n n) Definition (Markov Chain) 1 Finitely Many Outcomes. 2 Next State depends Only on Current State; gives rise to the matrix. 3 Always The Same (sketchy description, check the text). Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 9 / 13
10 Example M = v = [ ] Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 10 / 13
11 Inverse Matrix Definition The Identity Matrix I n n is It satisfies I M = M and N I = N for any M, N for which you can perform the multiplication. The inverse matrix to an n n matrix A is a matrix B such that A B = I. Alternatively B A = I. If one holds, both hold. If the inverse exists, it is unique. The inverse is written B 1. Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 11 / 13
12 Example of Inverse A = find its inverse if it exists Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 12 / 13
13 Encoding Encode the message To be or not to be. In numbers, 20, 15, 27, 2, 5, 27, 15, 18, 27, 14, 15, 20, 27, 20, 15, 27, 2, 5. A to Z are 1 to 26 and spaces are 27. You substitute, and break up into 3 1 column vectors. The encoded message is B v i. To decode, you break up into groups of three and make column vectors. Then apply B Dan Barbasch Math 1105 Chapter 2 and 10, October 23 Week of October 2 13 / 13
Chapter 2: Systems of Linear Equations and Matrices
Chapter 2: Systems of Linear Equations and Matrices 2.1 Systems Linear Equations: An Introduction Example Find the solution to the system of equations 2x y = 2 3x + 5y = 15 Solve first equation for y :
More information18.600: Lecture 32 Markov Chains
18.600: Lecture 32 Markov Chains Scott Sheffield MIT Outline Markov chains Examples Ergodicity and stationarity Outline Markov chains Examples Ergodicity and stationarity Markov chains Consider a sequence
More informationConstruction and Interpretation of Weather Station Models
Fall Semester Regents Earth Science John Dewey High School Mr. Klimetz Construction and Interpretation of Weather Station Models An Exercise in the Construction and Interpretation of Weather Station Models
More informationMarkov Chains. Chapter 16. Markov Chains - 1
Markov Chains Chapter 16 Markov Chains - 1 Why Study Markov Chains? Decision Analysis focuses on decision making in the face of uncertainty about one future event. However, many decisions need to consider
More information18.440: Lecture 33 Markov Chains
18.440: Lecture 33 Markov Chains Scott Sheffield MIT 1 Outline Markov chains Examples Ergodicity and stationarity 2 Outline Markov chains Examples Ergodicity and stationarity 3 Markov chains Consider a
More informationElementary Matrices. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics
Elementary Matrices MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Outline Today s discussion will focus on: elementary matrices and their properties, using elementary
More informationMarkov Model. Model representing the different resident states of a system, and the transitions between the different states
Markov Model Model representing the different resident states of a system, and the transitions between the different states (applicable to repairable, as well as non-repairable systems) System behavior
More informationMath Camp Notes: Linear Algebra II
Math Camp Notes: Linear Algebra II Eigenvalues Let A be a square matrix. An eigenvalue is a number λ which when subtracted from the diagonal elements of the matrix A creates a singular matrix. In other
More informationSpecial Mathematics. Tutorial 13. Markov chains
Tutorial 13 Markov chains "The future starts today, not tomorrow." Pope John Paul II A sequence of trials of an experiment is a nite Markov chain if: the outcome of each experiment is one of a nite set
More informationMATH 118 FINAL EXAM STUDY GUIDE
MATH 118 FINAL EXAM STUDY GUIDE Recommendations: 1. Take the Final Practice Exam and take note of questions 2. Use this study guide as you take the tests and cross off what you know well 3. Take the Practice
More information5.1 Introduction to Matrices
5.1 Introduction to Matrices Reminder: A matrix with m rows and n columns has size m x n. (This is also sometimes referred to as the order of the matrix.) The entry in the ith row and jth column of a matrix
More information18.175: Lecture 30 Markov chains
18.175: Lecture 30 Markov chains Scott Sheffield MIT Outline Review what you know about finite state Markov chains Finite state ergodicity and stationarity More general setup Outline Review what you know
More informationDefinition and Examples of DTMCs
Definition and Examples of DTMCs Natarajan Gautam Department of Industrial and Systems Engineering Texas A&M University 235A Zachry, College Station, TX 77843-3131 Email: gautam@tamuedu Phone: 979-845-5458
More informationMath 304 Handout: Linear algebra, graphs, and networks.
Math 30 Handout: Linear algebra, graphs, and networks. December, 006. GRAPHS AND ADJACENCY MATRICES. Definition. A graph is a collection of vertices connected by edges. A directed graph is a graph all
More informationLecture 5: Introduction to Markov Chains
Lecture 5: Introduction to Markov Chains Winfried Just Department of Mathematics, Ohio University January 24 26, 2018 weather.com light The weather is a stochastic process. For now we can assume that this
More informationA Mathematical Theory of Communication
A Mathematical Theory of Communication Ben Eggers Abstract This paper defines information-theoretic entropy and proves some elementary results about it. Notably, we prove that given a few basic assumptions
More informationCS188 Outline. CS 188: Artificial Intelligence. Today. Inference in Ghostbusters. Probability. We re done with Part I: Search and Planning!
CS188 Outline We re done with art I: Search and lanning! CS 188: Artificial Intelligence robability art II: robabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error
More informationWillmar Public Schools Curriculum Map
Subject Area Mathematics Senior High Course Name Advanced Algebra 2A (Prentice Hall Mathematics) Date April 2010 The Advanced Algebra 2A course parallels each other in content and time. The Advanced Algebra
More information8: Hidden Markov Models
8: Hidden Markov Models Machine Learning and Real-world Data Helen Yannakoudakis 1 Computer Laboratory University of Cambridge Lent 2018 1 Based on slides created by Simone Teufel So far we ve looked at
More informationMon Jan Matrix and linear transformations. Announcements: Warm-up Exercise:
Math 2270-004 Week 4 notes We will not necessarily finish the material from a given day's notes on that day. We may also add or subtract some material as the week progresses, but these notes represent
More information22m:033 Notes: 3.1 Introduction to Determinants
22m:033 Notes: 3. Introduction to Determinants Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman October 27, 2009 When does a 2 2 matrix have an inverse? ( ) a a If A =
More informationDiscrete Markov Chain. Theory and use
Discrete Markov Chain. Theory and use Andres Vallone PhD Student andres.vallone@predoc.uam.es 2016 Contents 1 Introduction 2 Concept and definition Examples Transitions Matrix Chains Classification 3 Empirical
More informationIntroduction to Matrices and Linear Systems Ch. 3
Introduction to Matrices and Linear Systems Ch. 3 Doreen De Leon Department of Mathematics, California State University, Fresno June, 5 Basic Matrix Concepts and Operations Section 3.4. Basic Matrix Concepts
More informationSolve the system of equations. 1) 1) SHORT ANSWER. Write the word or phrase that best completes each statement or answers the question.
Assignment Ch Name Solve the system of equations. ) ) x y z = - x - y z = x y z = x = -, y = -, z = ; (-, -, ) x =, y = -, z = -; (, -, -) x = -, y =, z = -; (-,, -) inconsistent ) The perimeter of a parking
More informationINVERSE OF A MATRIX [2.2] 8-1
INVERSE OF A MATRIX [2.2] 8-1 The inverse of a matrix: Introduction We have a mapping from R n to R n represented by a matrix A. Can we invert this mapping? i.e. can we find a matrix (call it B for now)
More informationBernoulli Counting Process with p=0.1
Stat 28 October 29, 21 Today: More Ch 7 (Sections 7.4 and part of 7.) Midterm will cover Ch 7 to section 7.4 Review session will be Nov. Exercises to try (answers in book): 7.1-, 7.2-3, 7.3-3, 7.4-7 Where
More informationCS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning
CS188 Outline We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more! Part III:
More information= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1
Properties of Markov Chains and Evaluation of Steady State Transition Matrix P ss V. Krishnan - 3/9/2 Property 1 Let X be a Markov Chain (MC) where X {X n : n, 1, }. The state space is E {i, j, k, }. The
More informationMath 180B Problem Set 3
Math 180B Problem Set 3 Problem 1. (Exercise 3.1.2) Solution. By the definition of conditional probabilities we have Pr{X 2 = 1, X 3 = 1 X 1 = 0} = Pr{X 3 = 1 X 2 = 1, X 1 = 0} Pr{X 2 = 1 X 1 = 0} = P
More informationIE 336 Seat # Name (clearly) Closed book. One page of hand-written notes, front and back. No calculator. 60 minutes.
Closed book. One page of hand-written notes, front and back. No calculator. 6 minutes. Cover page and four pages of exam. Fifteen questions. Each question is worth seven points. To receive full credit,
More informationCS 343: Artificial Intelligence
CS 343: Artificial Intelligence Probability Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188
More informationCoding Theory. Golay Codes
Coding Theory Massoud Malek Golay Codes The Golay codes are examples of perfect codes; they were discovered by the Swiss mathematician and information theorist, Marcel J. E. Golay in 1949. A binary Golay
More informationSHORT ANSWER. Write the word or phrase that best completes each statement or answers the question.
Assn.-.7 Name Perform the operation, if possible. ) Let A = 0 - - 0 and B = - 0 - -. Find AB. ) -6-9 ) Let A = - 5 and B = -5-7 - 6-8 - - - - - -. Find AB. ) 6-0 9 5-5 - -6-0 ) - 0 - - - 6-0 - 0 0 0 0
More informationElementary Matrices. which is obtained by multiplying the first row of I 3 by -1, and
Elementary Matrices In this special handout of material not contained in the text, we introduce the concept of elementary matrix. Elementary matrices are useful in several ways that will be shown in this
More information2 DISCRETE-TIME MARKOV CHAINS
1 2 DISCRETE-TIME MARKOV CHAINS 21 FUNDAMENTAL DEFINITIONS AND PROPERTIES From now on we will consider processes with a countable or finite state space S {0, 1, 2, } Definition 1 A discrete-time discrete-state
More informationCS 188: Artificial Intelligence Fall 2011
CS 188: Artificial Intelligence Fall 2011 Lecture 12: Probability 10/4/2011 Dan Klein UC Berkeley 1 Today Probability Random Variables Joint and Marginal Distributions Conditional Distribution Product
More informationExam 3 Review Math 118 Sections 1 and 2
Exam 3 Review Math 118 Sections 1 and 2 This exam will cover sections 5.3-5.6, 6.1-6.3 and 7.1-7.3 of the textbook. No books, notes, calculators or other aids are allowed on this exam. There is no time
More informationLesson 8: Variability in a Data Distribution
Classwork Example 1: Comparing Two Distributions Robert s family is planning to move to either New York City or San Francisco. Robert has a cousin in San Francisco and asked her how she likes living in
More informationPlease simplify your answers to the extent reasonable without a calculator. Show your work. Explain your answers, concisely.
Please simplify your answers to the extent reasonable without a calculator. Show your work. Explain your answers, concisely. 1. Consider a game which involves flipping a coin: winning $1 when it lands
More informationMon Feb Matrix algebra and matrix inverses. Announcements: Warm-up Exercise:
Math 2270-004 Week 5 notes We will not necessarily finish the material from a given day's notes on that day We may also add or subtract some material as the week progresses, but these notes represent an
More informationThe Boundary Problem: Markov Chain Solution
MATH 529 The Boundary Problem: Markov Chain Solution Consider a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability p, moves downward b units
More informationModule 8. Lecture 3: Markov chain
Lecture 3: Markov chain A Markov chain is a stochastic rocess having the roerty that the value of the rocess X t at time t, deends only on its value at time t-1, X t-1 and not on the sequence X t-2, X
More informationAlgebra: Linear UNIT 16 Equations Lesson Plan 1
1A UNIT 16 Equations Lesson Plan 1 Introduction to coding T: Today we're going to code and decode letters and numbers. OS 16.1 T: What can you see on the OHP? (A circle with letters in it) T: How is it
More informationWe Live in Exciting Times. CSCI-567: Machine Learning (Spring 2019) Outline. Outline. ACM (an international computing research society) has named
We Live in Exciting Times ACM (an international computing research society) has named CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Apr. 2, 2019 Yoshua Bengio,
More informationSTATISTICS 407 METHODS OF MULTIVARIATE ANALYSIS TOPICS
STATISTICS 407 METHODS OF MULTIVARIATE ANALYSIS TOPICS Principal Component Analysis (PCA): Reduce the, summarize the sources of variation in the data, transform the data into a new data set where the variables
More informationReinforcement Learning Wrap-up
Reinforcement Learning Wrap-up Slides courtesy of Dan Klein and Pieter Abbeel University of California, Berkeley [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationMath 120 Chapter 3 additional notes
Math 120 Chapter 3 additional notes MEASURES OF DISPERSION AND NORMAL DISTRIBUTION MEASURES OF DISPERSION In addition to knowing the measures of central tendency, it is often important to know how widely
More informationWe are here. Assembly Language. Processors Arithmetic Logic Units. Finite State Machines. Circuits Gates. Transistors
CSC258 Week 3 1 Logistics If you cannot login to MarkUs, email me your UTORID and name. Check lab marks on MarkUs, if it s recorded wrong, contact Larry within a week after the lab. Quiz 1 average: 86%
More informationFall Inverse of a matrix. Institute: UC San Diego. Authors: Alexander Knop
Fall 2017 Inverse of a matrix Authors: Alexander Knop Institute: UC San Diego Row-Column Rule If the product AB is defined, then the entry in row i and column j of AB is the sum of the products of corresponding
More informationThe probability of going from one state to another state on the next trial depends only on the present experiment and not on past history.
c Dr Oksana Shatalov, Fall 2010 1 9.1: Markov Chains DEFINITION 1. Markov process, or Markov Chain, is an experiment consisting of a finite number of stages in which the outcomes and associated probabilities
More informationToday s Outline. CS 362, Lecture 13. Matrix Chain Multiplication. Paranthesizing Matrices. Matrix Multiplication. Jared Saia University of New Mexico
Today s Outline CS 362, Lecture 13 Jared Saia University of New Mexico Matrix Multiplication 1 Matrix Chain Multiplication Paranthesizing Matrices Problem: We are given a sequence of n matrices, A 1, A
More informationChapter 2: Probability Part 1
Engineering Probability & Statistics (AGE 1150) Chapter 2: Probability Part 1 Dr. O. Phillips Agboola Sample Space (S) Experiment: is some procedure (or process) that we do and it results in an outcome.
More informationAnswers and Solutions to (Even Numbered) Suggested Exercises in Sections of Grimaldi s Discrete and Combinatorial Mathematics
Answers and Solutions to (Even Numbered) Suggested Exercises in Sections 6.5-6.9 of Grimaldi s Discrete and Combinatorial Mathematics Section 6.5 6.5.2. a. r = = + = c + e. So the error pattern is e =.
More informationReed-Muller Codes. These codes were discovered by Muller and the decoding by Reed in Code length: n = 2 m, Dimension: Minimum Distance
Reed-Muller Codes Ammar Abh-Hhdrohss Islamic University -Gaza ١ Reed-Muller Codes These codes were discovered by Muller and the decoding by Reed in 954. Code length: n = 2 m, Dimension: Minimum Distance
More informationWhere is the tropical zone? What are three biomes found in the tropical zone?
Name CHAPTER 3 Class Date Climate 2 The Tropics SECTION BEFORE YOU READ After you read this section, you should be able to answer these questions: Where is the tropical zone? What are three biomes found
More informationError Detection and Correction: Hamming Code; Reed-Muller Code
Error Detection and Correction: Hamming Code; Reed-Muller Code Greg Plaxton Theory in Programming Practice, Spring 2005 Department of Computer Science University of Texas at Austin Hamming Code: Motivation
More informationLecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity
5-859: Information Theory and Applications in TCS CMU: Spring 23 Lecture 8: Channel and source-channel coding theorems; BEC & linear codes February 7, 23 Lecturer: Venkatesan Guruswami Scribe: Dan Stahlke
More informationBayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction
15-0: Learning vs. Deduction Artificial Intelligence Programming Bayesian Learning Chris Brooks Department of Computer Science University of San Francisco So far, we ve seen two types of reasoning: Deductive
More informationOur Status in CSE 5522
Our Status in CSE 5522 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more!
More informationCut-Set Bound and Dependence Balance Bound
Cut-Set Bound and Dependence Balance Bound Lei Xiao lxiao@nd.edu 1 Date: 4 October, 2006 Reading: Elements of information theory by Cover and Thomas [1, Section 14.10], and the paper by Hekstra and Willems
More informationWorksheet 2 Problems
Technische Universität München WS 205/6 Lehrstuhl für Informatik V Scientific Computing Dr. T. Neckel 02..205/04..205 M.Sc. D. Jarema Worksheet 2 Problems Eigenvalue Problems and Algebraic Models (I) Exercise
More informationLinear Codes and Syndrome Decoding
Linear Codes and Syndrome Decoding These notes are intended to be used as supplementary reading to Sections 6.7 9 of Grimaldi s Discrete and Combinatorial Mathematics. The proofs of the theorems are left
More informationLinear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4
Linear Algebra Section. : LU Decomposition Section. : Permutations and transposes Wednesday, February 1th Math 01 Week # 1 The LU Decomposition We learned last time that we can factor a invertible matrix
More informationPrinciples of Food and Bioprocess Engineering (FS 231) Solutions to Example Problems on Mass Balance
Principles of Food and Bioprocess Engineering (FS 231) Solutions to Example Problems on Mass Balance 1. The mass balance equation for the system is: 2 + 3 = m This yields, m = 5 kg 2. The mass balance
More informationISM206 Lecture, May 12, 2005 Markov Chain
ISM206 Lecture, May 12, 2005 Markov Chain Instructor: Kevin Ross Scribe: Pritam Roy May 26, 2005 1 Outline of topics for the 10 AM lecture The topics are: Discrete Time Markov Chain Examples Chapman-Kolmogorov
More informationProblem Set 9 Due: In class Tuesday, Nov. 27 Late papers will be accepted until 12:00 on Thursday (at the beginning of class).
Math 3, Fall Jerry L. Kazdan Problem Set 9 Due In class Tuesday, Nov. 7 Late papers will be accepted until on Thursday (at the beginning of class).. Suppose that is an eigenvalue of an n n matrix A and
More informationLAB 3: THE SUN AND CLIMATE NAME: LAB PARTNER(S):
GEOG 101L PHYSICAL GEOGRAPHY LAB SAN DIEGO CITY COLLEGE SELKIN 1 LAB 3: THE SUN AND CLIMATE NAME: LAB PARTNER(S): The main objective of today s lab is for you to be able to visualize the sun s position
More informationDiscrete time Markov chains. Discrete Time Markov Chains, Definition and classification. Probability axioms and first results
Discrete time Markov chains Discrete Time Markov Chains, Definition and classification 1 1 Applied Mathematics and Computer Science 02407 Stochastic Processes 1, September 5 2017 Today: Short recap of
More information1 Introduction to information theory
1 Introduction to information theory 1.1 Introduction In this chapter we present some of the basic concepts of information theory. The situations we have in mind involve the exchange of information through
More informationSUPPLEMENTARY INFORMATION
SUPPLEMENTARY INFORMATION doi:10.1038/nature11875 Method for Encoding and Decoding Arbitrary Computer Files in DNA Fragments 1 Encoding 1.1: An arbitrary computer file is represented as a string S 0 of
More informationEigenvalues and Eigenvectors
Eigenvalues and Eigenvectors MAT 67L, Laboratory III Contents Instructions (1) Read this document. (2) The questions labeled Experiments are not graded, and should not be turned in. They are designed for
More informationNo class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.
Stationary Distributions Monday, September 28, 2015 2:02 PM No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1. Homework 1 due Friday, October 2 at 5 PM strongly
More informationCS1802 Week 11: Algorithms, Sums, Series, Induction
CS180 Discrete Structures Recitation Fall 017 Nov 11 - November 17, 017 CS180 Week 11: Algorithms, Sums, Series, Induction 1 Markov chain i. Boston has days which are either sunny or rainy and can be modeled
More informationDot Products, Transposes, and Orthogonal Projections
Dot Products, Transposes, and Orthogonal Projections David Jekel November 13, 2015 Properties of Dot Products Recall that the dot product or standard inner product on R n is given by x y = x 1 y 1 + +
More informationMarkov Chains and Transition Probabilities
Hinthada University Research Journal 215, Vol. 6, No. 1 3 Markov Chains and Transition Probabilities Ko Ko Oo Abstract Markov chain is widely applicable to the study of many real-world phenomene. We represent
More informationStochastic Problems. 1 Examples. 1.1 Neoclassical Growth Model with Stochastic Technology. 1.2 A Model of Job Search
Stochastic Problems References: SLP chapters 9, 10, 11; L&S chapters 2 and 6 1 Examples 1.1 Neoclassical Growth Model with Stochastic Technology Production function y = Af k where A is random Let A s t
More information4452 Mathematical Modeling Lecture 16: Markov Processes
Math Modeling Lecture 16: Markov Processes Page 1 4452 Mathematical Modeling Lecture 16: Markov Processes Introduction A stochastic model is one in which random effects are incorporated into the model.
More informationLet's contemplate a continuous-time limit of the Bernoulli process:
Mathematical Foundations of Markov Chains Thursday, September 17, 2015 2:04 PM Reading: Lawler Ch. 1 Homework 1 due Friday, October 2 at 5 PM. Office hours today are moved to 6-7 PM. Let's revisit the
More informationIntroduction to Information Retrieval
Introduction to Information Retrieval http://informationretrieval.org IIR 18: Latent Semantic Indexing Hinrich Schütze Center for Information and Language Processing, University of Munich 2013-07-10 1/43
More informationMATH Examination for the Module MATH-3152 (May 2009) Coding Theory. Time allowed: 2 hours. S = q
MATH-315201 This question paper consists of 6 printed pages, each of which is identified by the reference MATH-3152 Only approved basic scientific calculators may be used. c UNIVERSITY OF LEEDS Examination
More informationUnit 1 Matrices Notes Packet Period: Matrices
Algebra 2/Trig Unit 1 Matrices Notes Packet Name: Period: # Matrices (1) Page 203 204 #11 35 Odd (2) Page 203 204 #12 36 Even (3) Page 211 212 #4 6, 17 33 Odd (4) Page 211 212 #12 34 Even (5) Page 218
More informationDoing Physics with Random Numbers
Doing Physics with Random Numbers Andrew J. Schultz Department of Chemical and Biological Engineering University at Buffalo The State University of New York Concepts Random numbers can be used to measure
More informationComponents and change of basis
Math 20F Linear Algebra Lecture 16 1 Components and change of basis Slide 1 Review: Isomorphism Review: Components in a basis Unique representation in a basis Change of basis Review: Isomorphism Definition
More informationCoordinate Algebra Units 1 & 2 EOCT Review
Coordinate Algebra Units 1 & 2 EOCT Review Name: 1. A forest owned by the Jumbo Lumber Corporation contains 7,118 trees. If Jumbo Lumber cuts down 42 trees every day, which function can be used to find
More informationLesson 3: Networks and Matrix Arithmetic
Opening Exercise Suppose a subway line also connects the four cities. Here is the subway and bus line network. The bus routes connecting the cities are represented by solid lines, and the subway routes
More informationProbability. VCE Maths Methods - Unit 2 - Probability
Probability Probability Tree diagrams La ice diagrams Venn diagrams Karnough maps Probability tables Union & intersection rules Conditional probability Markov chains 1 Probability Probability is the mathematics
More informationLogic Design I (17.341) Fall Lecture Outline
Logic Design I (17.341) Fall 2011 Lecture Outline Class # 06 October 24, 2011 Dohn Bowden 1 Today s Lecture Administrative Main Logic Topic Homework 2 Course Admin 3 Administrative Admin for tonight Syllabus
More informationChapter 29 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M.
29 Markov Chains Definition of a Markov Chain Markov chains are one of the most fun tools of probability; they give a lot of power for very little effort. We will restrict ourselves to finite Markov chains.
More informationProbabilistic Robotics, Sebastian Thrun, 2005 Page 36, 37. <1 x = nf ) =1/ 3 <1 x = f ) =1 p(x = f ) = 0.01 p(x = nf ) = 0.
Probabilistic Robotics, Sebastian Thrun, 005 Page 36, 37 Method : Denote: z n : the sensor measurement in each time x: the state of the sensor f: the state of the sensor is faulty nf: the state of the
More informationProperties of Context-Free Languages
Properties of Context-Free Languages Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationINTRODUCTION TO INFORMATION & COMMUNICATION TECHNOLOGY LECTURE 8 : WEEK 8 CSC-110-T
INTRODUCTION TO INFORMATION & COMMUNICATION TECHNOLOGY LECTURE 8 : WEEK 8 CSC-110-T Credit : (2 + 1) / Week TEXT AND REF. BOOKS Text Book: Peter Norton (2011), Introduction to Computers, 7 /e, McGraw-Hill
More informationLecture 4: Proof of Shannon s theorem and an explicit code
CSE 533: Error-Correcting Codes (Autumn 006 Lecture 4: Proof of Shannon s theorem and an explicit code October 11, 006 Lecturer: Venkatesan Guruswami Scribe: Atri Rudra 1 Overview Last lecture we stated
More informationTMA Calculus 3. Lecture 21, April 3. Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013
TMA4115 - Calculus 3 Lecture 21, April 3 Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013 www.ntnu.no TMA4115 - Calculus 3, Lecture 21 Review of last week s lecture Last week
More informationThe cost/reward formula has two specific widely used applications:
Applications of Absorption Probability and Accumulated Cost/Reward Formulas for FDMC Friday, October 21, 2011 2:28 PM No class next week. No office hours either. Next class will be 11/01. The cost/reward
More informationMath 1314 Week #14 Notes
Math 3 Week # Notes Section 5.: A system of equations consists of two or more equations. A solution to a system of equations is a point that satisfies all the equations in the system. In this chapter,
More informationMon Feb Matrix inverses, the elementary matrix approach overview of skipped section 2.5. Announcements: Warm-up Exercise:
Math 2270-004 Week 6 notes We will not necessarily finish the material from a given day's notes on that day We may also add or subtract some material as the week progresses, but these notes represent an
More informationCS412: Lecture #17. Mridul Aanjaneya. March 19, 2015
CS: Lecture #7 Mridul Aanjaneya March 9, 5 Solving linear systems of equations Consider a lower triangular matrix L: l l l L = l 3 l 3 l 33 l n l nn A procedure similar to that for upper triangular systems
More informationTMA 4265 Stochastic Processes Semester project, fall 2014 Student number and
TMA 4265 Stochastic Processes Semester project, fall 2014 Student number 730631 and 732038 Exercise 1 We shall study a discrete Markov chain (MC) {X n } n=0 with state space S = {0, 1, 2, 3, 4, 5, 6}.
More information8: Hidden Markov Models
8: Hidden Markov Models Machine Learning and Real-world Data Simone Teufel and Ann Copestake Computer Laboratory University of Cambridge Lent 2017 Last session: catchup 1 Research ideas from sentiment
More informationMath 416, Spring 2010 Gram-Schmidt, the QR-factorization, Orthogonal Matrices March 4, 2010 GRAM-SCHMIDT, THE QR-FACTORIZATION, ORTHOGONAL MATRICES
Math 46, Spring 00 Gram-Schmidt, the QR-factorization, Orthogonal Matrices March 4, 00 GRAM-SCHMIDT, THE QR-FACTORIZATION, ORTHOGONAL MATRICES Recap Yesterday we talked about several new, important concepts
More information