Fast Multipole Methods: Fundamentals & Applications. Ramani Duraiswami Nail A. Gumerov

Size: px
Start display at page:

Download "Fast Multipole Methods: Fundamentals & Applications. Ramani Duraiswami Nail A. Gumerov"

Transcription

1 Fast Multipole Methods: Fundamentals & Applications Ramani Duraiswami Nail A. Gumerov

2 Week 1. Introduction. What are multipole methods and what is this course about. Problems from physics, mathematics, computer vision, statistics, etc. that can be solved with fast multipole methods. Historical review of the fast multipole method. Review of Complexity Homework.

3 Course Mechanics Background Needed Linear Algebra Matrices, Vectors, Linear Systems, LU Decomposition, Eigenvalue problem, SVD Numerical Analysis (Interpolation, Approximation, Finite Differences, Taylor Series, etc.) Programming Matlab, C/C++ and/or FORTRAN For particular applications in your domain you will need to know the background material there E.g., familiarity with partial differential equations or integral equations, or interpolation, or Participation essential!

4 Ideally Course Requirements you are an applied student in CS, engineering, physics or industry, who has a problem in mind and would like to explore how the FMM could be applied to it or, you are Applied Math/Scientific Computation student who wants to learn this important algorithm At the end of the course you will have a familiarity with Application of FMM to different problems Data structures for FMM, and analysis related to it. Course focus is on applying the methods to achieve solution. Theory will be developed as needed to proceed

5 Homework Will try to have it every week Account for 35 % of your grade Will not be excessive Essential for learning --- must do as opposed to just read. Homework handed out last class of a week. Due last class of next week Thanksgiving week no homework (you can turn in the previous homework after thanksgiving)

6 Projects & Exams There will be a final project that will require you to implement an FMM algorithm in a field of your choice, account for 25% of the grade. Project to be chosen latest by October 17. Implementation (reimplementation) of an FMM algorithm Use an existing FMM algorithm to solve an applied or mathematical problem you can discuss these with us, and we will help you choose one, and over the project If you already have a project in mind you can discuss it with us Exams intermediate exam worth 15%, week of October 16 final exam worth 25%. Finals Week

7 Class web & Mailing List Discussions, announcements Homework will be posted on the web No late homework (except by timely prior arrangement) Hardcopy (no web or submissions) Solutions will be posted after homework collected Links to papers etc.

8 Introductions addresses What are your interests? What do you want us to cover? Last three years outline is posted at the course web site

9 Scientific Computing Do science on the computer Accepted paradigm in many fields Computational physics Computational chemistry Computational biology Computational fluid dynamics Computational solid mechanics Stellar dynamics Molecular Modeling Machine Learning Data Mining Meteorology Computational Statistics Computational Electromagnetics Medical Imaging & Tomography Computational Acoustics Geology and Ocean Engineering Circuit design Photonics Nanotechnology Model the problem as best as we can Then represent the physical system discretely

10 Problem Size As problem size grows, in general fidelity of representation grows On the other hand cost of solving the problem Memory required to store the discrete representation also grow Way the cost creases with size is the memory or run time complexity of an algorithm Need a language to talk about fast and low memory algorithms Asymptotic complexity

11 Asymptotic Equivalence Let n the problem size Two algorithms with costs f(n) and g(n) are said to be equivalent: f(n) ~ g(n) f( n) lim = n g( n) 1

12 Little Oh Asymptotically smaller: f(n) = o(g(n)) f( n) lim = n g( n) 0

13 Big Oh Asymptotic Order of Growth: f(n) = O(g(n)) limsup f( n) n g( n) <

14 The Oh s If f = o(g) or f ~ g then f = O(g) lim = 0 lim = 1 lim <

15 The Oh s If f = o(g), then g O(f) f g lim = 0 lim = g f

16 Big Oh Equivalently, f(n) = O(g(n)) c,n 0 0 n n 0 f(n) c g(n)

17 Big Oh f(x) = O(g(x)) log scale blue stays below red c g(x) ln c f(x)

18 Complexity we will specify the function g(n) The most common complexities are O(1) - not proportional to any variable number, i.e. a fixed/constant amount of time O(N) - proportional to the size of N (this includes a loop to N and loops to constant multiples of N such as 0.5N, 2N, 2000N - no matter what that is, if you double N you expect (on average) the program to take twice as long) O(N 2 ) - proportional to N squared (you double N, you expect it to take four times longer - usually two nested loops both dependent on N). O(log N) - this is tricker to show - usually the result of binary splitting. O(N log N) this is usually caused by doing log N splits but also doing N amount of work at each "layer" of splitting.

19 Theta Same Order of Growth: f(n) = Θ(g(n)) f(n) = O(g(n)) and g(n) = O(f(n))

20 Log complexity If you half data at each stage then number of stages until you have a single item is given (roughly) by log 2 N. => binary search takes log 2 N time to find an item. All logs grow a constant amount apart (homework) So we normally just say log N not log 2 N. Log N grows very slowly

21 What is the Fast Multipole Method? An algorithm for achieving fast products of particular dense matrices with vectors Similar to the Fast Fourier Transform For the FFT, matrix entries are uniformly sampled complex exponentials For FMM, matrix entries are Derived from particular functions Functions satisfy known translation theorems Name is a bit unfortunate What is a multipole? We will return to this Why is this important?

22 Vectors and Matrices d dimensional column vector x and its transpose n d dimensional matrix M and its transpose M t

23 Matrix vector product s 1 = m 11 x 1 + m 12 x m 1d x d s 2 = m 21 x 1 + m 22 x m 2d x d s n = m n1 x 1 + m n2 x m nd x d Matrix vector product is identical to a sum s i = j=1d m ij x j So algorithm for fast matrix vector products is also a fast summation algorithm d products and sums per line N lines Total Nd products and Nd sums to calculate N entries If d N then matrix multiplication cost is O(N 2 ) Storage of the matrix entries is also O(N 2 )

24 Linear Systems Solve a system of equations Mx=s M is a N N matrix, x is a N vector, s is a N vector Direct solution (Gauss elimination, LU Decomposition, SVD, ) all need O(N 3 ) operations Iterative methods typically converge in k steps with each step needing a matrix vector multiply O(N 2 ) if properly designed, k<< N A fast matrix vector multiplication algorithm (O(N log N) operations) will speed all these algorithms

25 Is this important? Argument: Moore s law: Processor speed doubles every 18 months If we wait long enough the computer will get fast enough and let my inefficient algorithm tackle the problem Is this true? Yes for algorithms with same asymptotic complexity No!! For algorithms with different asymptotic complexity For a million variables, we would need about 16 generations of Moore s law before a O(N 2 ) algorithm was comparable with a O(N) algorithm Similarly, clever problem formulation can also achieve large savings.

26 Memory complexity Sometimes we are not able to fit a problem in available memory Don t care how long solution takes, just if we can solve it To store a N N matrix we need N 2 locations 1 GB RAM = =1,073,741,824 bytes => largest N is 32,768 Out of core algorithms copy partial results to disk, and keep only necessary part of the matrix in memory FMM allows reduction of memory complexity as well Elements of the matrix required for the product can be generated as needed

27 Parallel Processing Another way to tackle large problems is to throw more hardware (or more sophisticated hardware) at the problem If the work can be distributed across several processors (say P of them), then it can reduce the cost by some factor (ideally P). Usually requires reformulation of the solution algorithm (unless it is trivially parallel) Was hot in the 80s. Interest in this area again as purpose built parallel processing engines such as Graphical Processing Units, Multicore processors, FPGAs, etc. become widely available

28 The need for fast algorithms Grand challenge problems in large numbers of variables Simulation of physical systems Electromagnetics of complex systems Stellar clusters Protein folding Turbulence Learning theory Kernel methods Support Vector Machines Graphics and Vision Light scattering

29 General problems in these areas can be posed in terms of millions (10 6 ) or billions (10 9 ) of variables Recall Avogadro s numer ( molecules/mole

30 Dense and Sparse matrices Operation estimates are for dense matrices. Majority of elements of the matrix are non-zero However in many applications matrices are sparse A sparse matrix (or vector, or array) is one in which most of the elements are zero. If storage space is more important than access speed, it may be preferable to store a sparse matrix as a list of (index, value) pairs. For a given sparsity structure it may be possible to define a fast matrix-vector product/linear system algorithm

31 Can compute a x 1 a 1 x 1 0 a x 2 a 2 x a3 0 0 x3 = a3x a 4 0 x 4 a 4 x a 5 x 5 a 5 x 5 In 5 operations instead of 25 operations Sparse matrices are not our concern here

32 Structured matrices Fast algorithms have been found for many dense matrices Typically the matrices have some structure Definition: A dense matrix of order N N is called structured if its entries depend on only O(N) parameters. Most famous example the fast Fourier transform

33 Fourier Analysis Def.: mathematical techniques for breaking up a signal into its components (sinusoids) Jean Baptiste Joseph Fourier ( ) can represent any continuous periodic signal as a sum of sinusoidal waves = + +

34 Notation For f, periodic with period p Fourier transform f(t) -> F[k] F[k] = 1/p 0 p f (t) e -2 π i k t/p d t = 1/p 0 p f (t) cos(2 π k t/p) dt i / p 0 p f (t) sin(2 π k t/p) dt Inverse Fourier transform F[k] -> f(t) f(t) = Σ k in Z F[k] e 2 π i k t/p

35 DFT N-1 Σ p f (t) e -2 π i k t/p φ[n] e d t 0 n =0-2 π i k nh /p f(t) φ[n] = f(n h) h

36 DFT and its inverse for periodic discrete data Φ[k] = N-1 Σ n =0 φ[n] e -2 π i k n h /p p = N h N-1 = Σ n=0 φ[n] e -2 π i k n / N This is automatically periodic in k with period N Inverse is like Fourier series, but with only p terms

37 Discrete time Numerical Fourier Analysis DFT is really just a matrix multiplication! F [m] = 1/Ν F[0] F[1] F[2]... F[N-1] = Freq. index N-1 Σ k =0 e -2 π i k m/n f[k] time index f[0] f[1] f[2]... f[n-1] F = F f N

38 Fourier Matrices

39 i w 4 = Primitive Roots of Unity Example: The complex number e 2pi/n is a primitive n-th root of unity, where w 3 1 A number ω is a primitive n-th root of unity, for n>1, if ω n = 1 The numbers 1, ω, ω 2,, ω n-1 are all distinct 2πi w 2 1 n Imaginary 45 o cos(π/4) w 1 =cos(π/4) + i w 8 Real Check: if properties are satisfied 1. ω = e 1 n 2 i n π n 2πi 2. ω = e = e = cos2π + isin 2π = 1 n 1 jp 0 j 3. S= ω = ω + ω 2 j 3 j + ω + ω j( n 1) ω = 0 p= 0 w 5 w 6 w 7 n complex roots of unity equally spaced around the circle of unit radius centered at the origin of the complex plane.

40 Roots of Unity: Properties Property 1: Let ω be the principal n th root of unity. If n > 0, then ω n/2 = -1. Proof: ω = e 2π i / n ω n/2 = e π i = -1. (Euler's formula) Reflective Property: Corollary: ω k+n/2 = -ω k. Property 2: Let n > 0 be even, and let ω and ν be the principal n th and (n/2) th roots of unity. Then (ω k ) 2 = ν k. Proof:(ω k ) 2 = e (2k)2π i / n = e (k) 2π i / (n / 2) = ν k. Reduction Property: If ω is a primitive (2n)-th root of unity, then ω 2 is a primitive n-th root of unity.

41 L3: Let n > 0 be even. Then, the squares of the n complex n th roots of unity are the n/2 complex (n/2) th roots of unity. Proof: If we square all of the n th roots of unity, then each (n/2) th root is obtained exactly twice since: L1 ω k + n / 2 = - ω k thus, (ω k + n / 2 ) 2 = (ω k ) 2 L2 both of these = ν k ω k + n / 2 and ω k have the same square Inverse Property: If ω is a primitive root of unity, then ω -1 =ω n-1 Proof: ωω n-1 =ω n =1

42 Fast Fourier Transform Presented by Cooley and Tukey in 1965, but invented several times, including by Gauss (1809) and Danielson & Lanczos (1948) Danielson Lanczos lemma

43 So far we have seen what happens on the right hand side How about the left hand side? When we split the sums in two we have two sets of sums with N/2 quantities for N points. So the complexity is N 2 /2 + N 2 /2 =N 2 So there is no improvement Need to reduce the sums on the right hand side as well We need to reduce the number of sums computed from 2N to a lower number Notice that the values corresponding to k and k+n/2 will be the same. The transforms F ek and F ok are periodic in k with length N/2. So we need only compute half of them!

44 FFT So DFT of order N can be expressed as sum of two DFTs of order N/2 evaluated at N/2 points Does this improve the complexity? Yes (N/2) 2 +(N/2) 2 =N 2 /2< N 2 But we are not done. Can apply the lemma recursively Finally we have a set of one point transforms One point transform is identity

45 FFT Algorithm FFT (n, a 0, a 1, a 2,..., a n-1 ) if (n == 1) // n is a power of 2 return a 0 ω e 2π i / n (e 0,e 1,e 2,...,e n/2-1 ) FFT(n/2, a 0,a 2,a 4,...,a n-2 ) (d 0,d 1,d 2,...,d n/2-1 ) FFT(n/2, a 1,a 3,a 5,...,a n-1 ) for k = 0 to n/2-1 y k e k + ω k d k y k+n/2 e k - ω k d k O(n) complex multiplies if we pre-compute ω k. return (y 0,y 1,y 2,...,y n-1 ) T ( n) = 2T ( n / 2) + O( n) T ( n) = O( nlogn)

46 Complexity Each F k is a sum of log 2 N transforms and (factors) There are N F k s So the algorithm is O(N log 2 N) This is a recursive algorithm

47 function y = ffttx(x) %FFTTX Textbook Fast Finite Fourier Transform. % FFTTX(X) computes the same finite Fourier transform as FFT(X). % The code uses a recursive divide and conquer algorithm for % even order and matrix-vector multiplication for odd order. % If length(x) is m*p where m is odd and p is a power of 2, the % computational complexity of this approach is O(m^2)*O(p*log2(p)). x = x(:); n = length(x); omega = exp(-2*pi*i/n); FFTtx if rem(n,2) == 0 % Recursive divide and conquer k = (0:n/2-1)'; w = omega.^ k; u = ffttx(x(1:2:n-1)); v = w.*ffttx(x(2:2:n)); y = [u+v; u-v]; else % The Fourier matrix. j = 0:n-1; k = j'; F = omega.^ (k*j); y = F*x; end

48 Scrambled Output of the FFT a 0, a 1, a 2, a 3, a 4, a 5, a 6, a 7 a 0, a 2, a 4, a 6 a 1, a 3, a 5, a 7 a 0, a 4 a 2, a 6 a 1, a 5 a 3, a 7 a 0 a 4 a 2 a 6 a 1 a 5 a 3 a 7 "bit-reversed" order

49 FFT and IFFT The FFT has revolutionized many applications by reducing the complexity by a factor of almost n Can relate many other matrices to the Fourier Matrix

50 Circulant Matrices Toeplitz Matrices Hankel Matrices Vandermonde Matrices

51 Modern signal processing very strongly based on the FFT One of the defining inventions of the 20 th century

52 Fast Multipole Methods (FMM) Introduced by Rokhlin & Greengard in 1987 Called one of the 10 most significant advances in computing of the 20 th century Speeds up matrix-vector products (sums) of a particular type Above sum requires O(MN) operations. For a given precision ε the FMM achieves the evaluation in O(M+N) operations.

53 Can accelerate matrix vector products Convert O(N 2 ) to O(N log N) However, can also accelerate linear system solution Use iterative methods that take k steps and use a matrix vector product at each step Convert O(N 3 ) to O(kN log N)

54 A very simple algorithm Not FMM, but has some key ideas Consider S(x i )= j=1n α j (x i y j ) 2 i=1,,m Naïve way to evaluate the sum will require MN operations Instead can write the sum as S(x i )=( j=1n α j )x i2 + ( j=1n α j y j2 ) -2x i ( j=1n α j y j ) Can evaluate each bracketed sum over j and evaluate an expression of the type S(x i )=β x i2 + γ -2x i δ Requires O(M+N) operations Key idea use of analytical manipulation of series to achieve faster summation

55 Approximate evaluation FMM introduces another key idea or philosophy In scientific computing we almost never seek exact answers At best, exact means to machine precision So instead of solving the problem we can solve a nearby problem that gives almost the same answer If this nearby problem is much easier to solve, and we can bound the error analytically we are done. In the case of the FMM Manipulate series to achieve approximate evaluation Use analytical expression to bound the error FFT is exact FMM can be arbitrarily accurate

56 Some FMM algorithms Molecular and stellar dynamics Computation of force fields and dynamics Interpolation with Radial Basis Functions Solution of acoustical scattering problems Helmholtz Equation Electromagnetic Wave scattering Maxwell s equations Fluid Mechanics: Potential flow, vortex flow Laplace/Poisson equations Fast nonuniform Fourier transform

57 Applications I Interpolation Given a scattered data set with points and values {x i, f i } Build a representation of the function f(x) That satisfies f(x i )=f i Can be evaluated at new points One approach use radial-basis functions f(x)= in α i R(x-x i ) + p(x) f j = in α i R(x j -x i ) + p(x j ) Two problems Determining α i Knowing α i determine the product at many new points x j Both can be solved via FMM (Cherrie et al, 2001; Gumerov & Duraiswami, 2006)

58 RBF interpolation Applications 2 Cherrie et al 2001 Original unstructured data Gumerov & Duraiswami, 2006 RBF interpolation to regular spatial grid

59 Applications 3 Sound scattering off rooms and bodies Need to know the scattering properties of the head and body (our interest) 2 2 P+ k P = py ( ) Gxyk (, ; ) ( ) ( ) = (, ; ) ( ) Γ ( ) y G xy, Cxpx Gxyk py d ny ny Γy 0 P + i σ P = g n P lim r ikp = 0 r r ik e x y = 4π x y

60 BEM/FMM computations of acoustic scattering from complex shape objects Incident plane wave exp(ikz) Mesh: elements vertices db Hz (kd=0.96) 2.5 khz (kd=9.6) 25 khz (kd=96) -9

61 Sound Pressure kd=0.96

62 Sound Pressure kd=9.6

63 Sound Pressure kd=96

64 Potential 125 randomly oriented ellipsoids vertices and elements (1016 vertices and 2028 elements per ellipsoid)

65 1000 randomly oriented ellipsoids vertices and elements (488 vertices and 972 elements per ellipsoid) Potential

66

67 EM wave scattering Similar to acoustic scattering Send waves and measure scattered waves Attempt to figure out object from the measured waves Need to know Radar cross-section Many applications Light scattering Radar Antenna design. Darve 2001

68 Molecular and stellar dynamics Many particles distributed in space Particles exert a force on each other Simplest case force obeys an inverse-square law (gravity, coulombic interaction) d x dt = ( x x ) 2 N i i j = F, 2 i Fi = qq i j 3 j 1 xi x j j i

69 Fluid mechanics Incompressible Navier Stokes Equation u = 0 u ρ + t + 2 ( u ) u p = µ u u = φ + A Laplace equation for potential and Poisson equation for vorticity Solved via particle methods

70 Vortex Methods

71 Applications Digital Signal Processing: analyzing and manipulating real-world signals using a computer Toys and consumer electronics Speech recognition Audio/video compression Medical imaging Communications Radar

Fast Multipole Methods: Fundamentals & Applications. Ramani Duraiswami Nail A. Gumerov

Fast Multipole Methods: Fundamentals & Applications. Ramani Duraiswami Nail A. Gumerov Fast Multipole Methods: Fundamentals & Applications Ramani Duraiswami Nail A. Gumerov Week 1. Introduction. What are multipole methods and what is this course about. Problems from physics, mathematics,

More information

Course Requirements. Course Mechanics. Projects & Exams. Homework. Week 1. Introduction. Fast Multipole Methods: Fundamentals & Applications

Course Requirements. Course Mechanics. Projects & Exams. Homework. Week 1. Introduction. Fast Multipole Methods: Fundamentals & Applications Week 1. Introduction. Fast Multipole Methods: Fundamentals & Applications Ramani Duraiswami Nail A. Gumerov What are multipole methods and what is this course about. Problems from phsics, mathematics,

More information

Computational Methods CMSC/AMSC/MAPL 460

Computational Methods CMSC/AMSC/MAPL 460 Computational Methods CMSC/AMSC/MAPL 460 Fourier transform Balaji Vasan Srinivasan Dept of Computer Science Several slides from Prof Healy s course at UMD Last time: Fourier analysis F(t) = A 0 /2 + A

More information

Fast Multipole Methods

Fast Multipole Methods An Introduction to Fast Multipole Methods Ramani Duraiswami Institute for Advanced Computer Studies University of Maryland, College Park http://www.umiacs.umd.edu/~ramani Joint work with Nail A. Gumerov

More information

Design and Analysis of Algorithms

Design and Analysis of Algorithms Design and Analysis of Algorithms CSE 5311 Lecture 5 Divide and Conquer: Fast Fourier Transform Junzhou Huang, Ph.D. Department of Computer Science and Engineering CSE5311 Design and Analysis of Algorithms

More information

Computational Methods CMSC/AMSC/MAPL 460. Fourier transform

Computational Methods CMSC/AMSC/MAPL 460. Fourier transform Computational Methods CMSC/AMSC/MAPL 460 Fourier transform Ramani Duraiswami, Dept. of Computer Science Several slides from Prof. Healy s course at UMD Fourier Methods Fourier analysis ( harmonic analysis

More information

5.6 Convolution and FFT

5.6 Convolution and FFT 5.6 Convolution and FFT Fast Fourier Transform: Applications Applications. Optics, acoustics, quantum physics, telecommunications, control systems, signal processing, speech recognition, data compression,

More information

Fast Multipole Methods for The Laplace Equation. Outline

Fast Multipole Methods for The Laplace Equation. Outline Fast Multipole Methods for The Laplace Equation Ramani Duraiswami Nail Gumerov Outline 3D Laplace equation and Coulomb potentials Multipole and local expansions Special functions Legendre polynomials Associated

More information

CS 4407 Algorithms Lecture 2: Iterative and Divide and Conquer Algorithms

CS 4407 Algorithms Lecture 2: Iterative and Divide and Conquer Algorithms CS 4407 Algorithms Lecture 2: Iterative and Divide and Conquer Algorithms Prof. Gregory Provan Department of Computer Science University College Cork 1 Lecture Outline CS 4407, Algorithms Growth Functions

More information

Partial Differential Equations. Examples of PDEs

Partial Differential Equations. Examples of PDEs Partial Differential Equations Almost all the elementary and numerous advanced parts of theoretical physics are formulated in terms of differential equations (DE). Newton s Laws Maxwell equations Schrodinger

More information

Algorithm Design and Analysis

Algorithm Design and Analysis Algorithm Design and Analysis LECTURE 14 Divide and Conquer Fast Fourier Transform Sofya Raskhodnikova 10/7/2016 S. Raskhodnikova; based on slides by K. Wayne. 5.6 Convolution and FFT Fast Fourier Transform:

More information

A Crash Course on Fast Multipole Method (FMM)

A Crash Course on Fast Multipole Method (FMM) A Crash Course on Fast Multipole Method (FMM) Hanliang Guo Kanso Biodynamical System Lab, 2018 April What is FMM? A fast algorithm to study N body interactions (Gravitational, Coulombic interactions )

More information

Fast Multipole Methods for Incompressible Flow Simulation

Fast Multipole Methods for Incompressible Flow Simulation Fast Multipole Methods for Incompressible Flow Simulation Nail A. Gumerov & Ramani Duraiswami Institute for Advanced Computer Studies University of Maryland, College Park Support of NSF awards 0086075

More information

How to Multiply. 5.5 Integer Multiplication. Complex Multiplication. Integer Arithmetic. Complex multiplication. (a + bi) (c + di) = x + yi.

How to Multiply. 5.5 Integer Multiplication. Complex Multiplication. Integer Arithmetic. Complex multiplication. (a + bi) (c + di) = x + yi. How to ultiply Slides by Kevin Wayne. Copyright 5 Pearson-Addison Wesley. All rights reserved. integers, matrices, and polynomials Complex ultiplication Complex multiplication. a + bi) c + di) = x + yi.

More information

CS711008Z Algorithm Design and Analysis

CS711008Z Algorithm Design and Analysis CS711008Z Algorithm Design and Analysis Lecture 5 FFT and Divide and Conquer Dongbo Bu Institute of Computing Technology Chinese Academy of Sciences, Beijing, China 1 / 56 Outline DFT: evaluate a polynomial

More information

Lecture 5. The Digital Fourier Transform. (Based, in part, on The Scientist and Engineer's Guide to Digital Signal Processing by Steven Smith)

Lecture 5. The Digital Fourier Transform. (Based, in part, on The Scientist and Engineer's Guide to Digital Signal Processing by Steven Smith) Lecture 5 The Digital Fourier Transform (Based, in part, on The Scientist and Engineer's Guide to Digital Signal Processing by Steven Smith) 1 -. 8 -. 6 -. 4 -. 2-1 -. 8 -. 6 -. 4 -. 2 -. 2. 4. 6. 8 1

More information

Analysis of Algorithms

Analysis of Algorithms October 1, 2015 Analysis of Algorithms CS 141, Fall 2015 1 Analysis of Algorithms: Issues Correctness/Optimality Running time ( time complexity ) Memory requirements ( space complexity ) Power I/O utilization

More information

Sequential Fast Fourier Transform (PSC ) Sequential FFT

Sequential Fast Fourier Transform (PSC ) Sequential FFT Sequential Fast Fourier Transform (PSC 3.1 3.2) 1 / 18 Applications of Fourier analysis Fourier analysis studies the decomposition of functions into their frequency components. Piano Concerto no. 9 by

More information

CS 542G: Conditioning, BLAS, LU Factorization

CS 542G: Conditioning, BLAS, LU Factorization CS 542G: Conditioning, BLAS, LU Factorization Robert Bridson September 22, 2008 1 Why some RBF Kernel Functions Fail We derived some sensible RBF kernel functions, like φ(r) = r 2 log r, from basic principles

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

Periodic motions. Periodic motions are known since the beginning of mankind: Motion of planets around the Sun; Pendulum; And many more...

Periodic motions. Periodic motions are known since the beginning of mankind: Motion of planets around the Sun; Pendulum; And many more... Periodic motions Periodic motions are known since the beginning of mankind: Motion of planets around the Sun; Pendulum; And many more... Periodic motions There are several quantities which describe a periodic

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Logistics Notes for 2016-08-26 1. Our enrollment is at 50, and there are still a few students who want to get in. We only have 50 seats in the room, and I cannot increase the cap further. So if you are

More information

Asymptotic Notation. such that t(n) cf(n) for all n n 0. for some positive real constant c and integer threshold n 0

Asymptotic Notation. such that t(n) cf(n) for all n n 0. for some positive real constant c and integer threshold n 0 Asymptotic Notation Asymptotic notation deals with the behaviour of a function in the limit, that is, for sufficiently large values of its parameter. Often, when analysing the run time of an algorithm,

More information

CS S Lecture 5 January 29, 2019

CS S Lecture 5 January 29, 2019 CS 6363.005.19S Lecture 5 January 29, 2019 Main topics are #divide-and-conquer with #fast_fourier_transforms. Prelude Homework 1 is due Tuesday, February 5th. I hope you ve at least looked at it by now!

More information

5 + 9(10) + 3(100) + 0(1000) + 2(10000) =

5 + 9(10) + 3(100) + 0(1000) + 2(10000) = Chapter 5 Analyzing Algorithms So far we have been proving statements about databases, mathematics and arithmetic, or sequences of numbers. Though these types of statements are common in computer science,

More information

DIVIDE AND CONQUER II

DIVIDE AND CONQUER II DIVIDE AND CONQUER II master theorem integer multiplication matrix multiplication convolution and FFT Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley http://www.cs.princeton.edu/~wayne/kleinberg-tardos

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

Asymptotic Analysis. Thomas A. Anastasio. January 7, 2004

Asymptotic Analysis. Thomas A. Anastasio. January 7, 2004 Asymptotic Analysis Thomas A. Anastasio January 7, 004 1 Introduction As a programmer, you often have a choice of data structures and algorithms. Choosing the best one for a particular job involves, among

More information

Great Theoretical Ideas in Computer Science. Lecture 9: Introduction to Computational Complexity

Great Theoretical Ideas in Computer Science. Lecture 9: Introduction to Computational Complexity 15-251 Great Theoretical Ideas in Computer Science Lecture 9: Introduction to Computational Complexity February 14th, 2017 Poll What is the running time of this algorithm? Choose the tightest bound. def

More information

ESE 250: Digital Audio Basics. Week 4 February 5, The Frequency Domain. ESE Spring'13 DeHon, Kod, Kadric, Wilson-Shah

ESE 250: Digital Audio Basics. Week 4 February 5, The Frequency Domain. ESE Spring'13 DeHon, Kod, Kadric, Wilson-Shah ESE 250: Digital Audio Basics Week 4 February 5, 2013 The Frequency Domain 1 Course Map 2 Musical Representation With this compact notation Could communicate a sound to pianist Much more compact than 44KHz

More information

This chapter covers asymptotic analysis of function growth and big-o notation.

This chapter covers asymptotic analysis of function growth and big-o notation. Chapter 14 Big-O This chapter covers asymptotic analysis of function growth and big-o notation. 14.1 Running times of programs An important aspect of designing a computer programs is figuring out how well

More information

An analogy from Calculus: limits

An analogy from Calculus: limits COMP 250 Fall 2018 35 - big O Nov. 30, 2018 We have seen several algorithms in the course, and we have loosely characterized their runtimes in terms of the size n of the input. We say that the algorithm

More information

CS 4407 Algorithms Lecture 2: Growth Functions

CS 4407 Algorithms Lecture 2: Growth Functions CS 4407 Algorithms Lecture 2: Growth Functions Prof. Gregory Provan Department of Computer Science University College Cork 1 Lecture Outline Growth Functions Mathematical specification of growth functions

More information

Growth of Functions (CLRS 2.3,3)

Growth of Functions (CLRS 2.3,3) Growth of Functions (CLRS 2.3,3) 1 Review Last time we discussed running time of algorithms and introduced the RAM model of computation. Best-case running time: the shortest running time for any input

More information

Analysis of Algorithms

Analysis of Algorithms Presentation for use with the textbook Data Structures and Algorithms in Java, 6th edition, by M. T. Goodrich, R. Tamassia, and M. H. Goldwasser, Wiley, 2014 Analysis of Algorithms Input Algorithm Analysis

More information

Algorithms and Their Complexity

Algorithms and Their Complexity CSCE 222 Discrete Structures for Computing David Kebo Houngninou Algorithms and Their Complexity Chapter 3 Algorithm An algorithm is a finite sequence of steps that solves a problem. Computational complexity

More information

Great Theoretical Ideas in Computer Science. Lecture 7: Introduction to Computational Complexity

Great Theoretical Ideas in Computer Science. Lecture 7: Introduction to Computational Complexity 15-251 Great Theoretical Ideas in Computer Science Lecture 7: Introduction to Computational Complexity September 20th, 2016 What have we done so far? What will we do next? What have we done so far? > Introduction

More information

Transforms and Orthogonal Bases

Transforms and Orthogonal Bases Orthogonal Bases Transforms and Orthogonal Bases We now turn back to linear algebra to understand transforms, which map signals between different domains Recall that signals can be interpreted as vectors

More information

Sparse Linear Systems. Iterative Methods for Sparse Linear Systems. Motivation for Studying Sparse Linear Systems. Partial Differential Equations

Sparse Linear Systems. Iterative Methods for Sparse Linear Systems. Motivation for Studying Sparse Linear Systems. Partial Differential Equations Sparse Linear Systems Iterative Methods for Sparse Linear Systems Matrix Computations and Applications, Lecture C11 Fredrik Bengzon, Robert Söderlund We consider the problem of solving the linear system

More information

Big O 2/14/13. Administrative. Does it terminate? David Kauchak cs302 Spring 2013

Big O 2/14/13. Administrative. Does it terminate? David Kauchak cs302 Spring 2013 /4/3 Administrative Big O David Kauchak cs3 Spring 3 l Assignment : how d it go? l Assignment : out soon l CLRS code? l Videos Insertion-sort Insertion-sort Does it terminate? /4/3 Insertion-sort Loop

More information

Cpt S 223. School of EECS, WSU

Cpt S 223. School of EECS, WSU Algorithm Analysis 1 Purpose Why bother analyzing code; isn t getting it to work enough? Estimate time and memory in the average case and worst case Identify bottlenecks, i.e., where to reduce time Compare

More information

Fast multipole method

Fast multipole method 203/0/ page Chapter 4 Fast multipole method Originally designed for particle simulations One of the most well nown way to tae advantage of the low-ran properties Applications include Simulation of physical

More information

Module 1: Analyzing the Efficiency of Algorithms

Module 1: Analyzing the Efficiency of Algorithms Module 1: Analyzing the Efficiency of Algorithms Dr. Natarajan Meghanathan Associate Professor of Computer Science Jackson State University Jackson, MS 39217 E-mail: natarajan.meghanathan@jsums.edu Based

More information

CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis. Ruth Anderson Winter 2019

CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis. Ruth Anderson Winter 2019 CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis Ruth Anderson Winter 2019 Today Algorithm Analysis What do we care about? How to compare two algorithms Analyzing Code Asymptotic Analysis

More information

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

Math 471 (Numerical methods) Chapter 3 (second half). System of equations Math 47 (Numerical methods) Chapter 3 (second half). System of equations Overlap 3.5 3.8 of Bradie 3.5 LU factorization w/o pivoting. Motivation: ( ) A I Gaussian Elimination (U L ) where U is upper triangular

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

FastMMLib: a generic Fast Multipole Method library. Eric DARRIGRAND. IRMAR Université de Rennes 1

FastMMLib: a generic Fast Multipole Method library. Eric DARRIGRAND. IRMAR Université de Rennes 1 FastMMLib: a generic Fast Multipole Method library Eric DARRIGRAND joint work with Yvon LAFRANCHE IRMAR Université de Rennes 1 Outline Introduction to FMM Historically SVD principle A first example of

More information

Anouncements. Assignment 3 has been posted!

Anouncements. Assignment 3 has been posted! Anouncements Assignment 3 has been posted! FFTs in Graphics and Vision Correlation of Spherical Functions Outline Math Review Spherical Correlation Review Dimensionality: Given a complex n-dimensional

More information

CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis. Ruth Anderson Winter 2018

CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis. Ruth Anderson Winter 2018 CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis Ruth Anderson Winter 2018 Today Algorithm Analysis What do we care about? How to compare two algorithms Analyzing Code Asymptotic Analysis

More information

ABSTRACT MATRICES WITH APPLICATIONS TO THE FAST MULTIPOLE METHOD. Zhihui Tang, Doctor of Philosophy, 2004

ABSTRACT MATRICES WITH APPLICATIONS TO THE FAST MULTIPOLE METHOD. Zhihui Tang, Doctor of Philosophy, 2004 ABSTRACT Title of Dissertation: FAST TRANSFORMS BASED ON STRUCTURED MATRICES WITH APPLICATIONS TO THE FAST MULTIPOLE METHOD Zhihui Tang, Doctor of Philosophy, 2004 Dissertation directed by: Professor Ramani

More information

CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis. Ruth Anderson Winter 2018

CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis. Ruth Anderson Winter 2018 CSE332: Data Structures & Parallelism Lecture 2: Algorithm Analysis Ruth Anderson Winter 2018 Today Algorithm Analysis What do we care about? How to compare two algorithms Analyzing Code Asymptotic Analysis

More information

Divide and Conquer. Maximum/minimum. Median finding. CS125 Lecture 4 Fall 2016

Divide and Conquer. Maximum/minimum. Median finding. CS125 Lecture 4 Fall 2016 CS125 Lecture 4 Fall 2016 Divide and Conquer We have seen one general paradigm for finding algorithms: the greedy approach. We now consider another general paradigm, known as divide and conquer. We have

More information

Analysis of Algorithm Efficiency. Dr. Yingwu Zhu

Analysis of Algorithm Efficiency. Dr. Yingwu Zhu Analysis of Algorithm Efficiency Dr. Yingwu Zhu Measure Algorithm Efficiency Time efficiency How fast the algorithm runs; amount of time required to accomplish the task Our focus! Space efficiency Amount

More information

FFTs in Graphics and Vision. Homogenous Polynomials and Irreducible Representations

FFTs in Graphics and Vision. Homogenous Polynomials and Irreducible Representations FFTs in Graphics and Vision Homogenous Polynomials and Irreducible Representations 1 Outline The 2π Term in Assignment 1 Homogenous Polynomials Representations of Functions on the Unit-Circle Sub-Representations

More information

Advanced Algorithmics (6EAP)

Advanced Algorithmics (6EAP) Advanced Algorithmics (6EAP) MTAT.03.238 Order of growth maths Jaak Vilo 2017 fall Jaak Vilo 1 Program execution on input of size n How many steps/cycles a processor would need to do How to relate algorithm

More information

When we use asymptotic notation within an expression, the asymptotic notation is shorthand for an unspecified function satisfying the relation:

When we use asymptotic notation within an expression, the asymptotic notation is shorthand for an unspecified function satisfying the relation: CS 124 Section #1 Big-Oh, the Master Theorem, and MergeSort 1/29/2018 1 Big-Oh Notation 1.1 Definition Big-Oh notation is a way to describe the rate of growth of functions. In CS, we use it to describe

More information

CS168: The Modern Algorithmic Toolbox Lecture #11: The Fourier Transform and Convolution

CS168: The Modern Algorithmic Toolbox Lecture #11: The Fourier Transform and Convolution CS168: The Modern Algorithmic Toolbox Lecture #11: The Fourier Transform and Convolution Tim Roughgarden & Gregory Valiant May 8, 2015 1 Intro Thus far, we have seen a number of different approaches to

More information

MA008/MIIZ01 Design and Analysis of Algorithms Lecture Notes 2

MA008/MIIZ01 Design and Analysis of Algorithms Lecture Notes 2 MA008 p.1/36 MA008/MIIZ01 Design and Analysis of Algorithms Lecture Notes 2 Dr. Markus Hagenbuchner markus@uow.edu.au. MA008 p.2/36 Content of lecture 2 Examples Review data structures Data types vs. data

More information

The Fast Fourier Transform. Andreas Klappenecker

The Fast Fourier Transform. Andreas Klappenecker The Fast Fourier Transform Andreas Klappenecker Motivation There are few algorithms that had more impact on modern society than the fast Fourier transform and its relatives. The applications of the fast

More information

Tutorial 2 - Learning about the Discrete Fourier Transform

Tutorial 2 - Learning about the Discrete Fourier Transform Tutorial - Learning about the Discrete Fourier Transform This tutorial will be about the Discrete Fourier Transform basis, or the DFT basis in short. What is a basis? If we google define basis, we get:

More information

MATH20602 Numerical Analysis 1

MATH20602 Numerical Analysis 1 M\cr NA Manchester Numerical Analysis MATH20602 Numerical Analysis 1 Martin Lotz School of Mathematics The University of Manchester Manchester, January 27, 2014 Outline General Course Information Introduction

More information

CHAPTER 11. A Revision. 1. The Computers and Numbers therein

CHAPTER 11. A Revision. 1. The Computers and Numbers therein CHAPTER A Revision. The Computers and Numbers therein Traditional computer science begins with a finite alphabet. By stringing elements of the alphabet one after another, one obtains strings. A set of

More information

NUMERICAL MATHEMATICS & COMPUTING 7th Edition

NUMERICAL MATHEMATICS & COMPUTING 7th Edition NUMERICAL MATHEMATICS & COMPUTING 7th Edition Ward Cheney/David Kincaid c UT Austin Engage Learning: Thomson-Brooks/Cole wwwengagecom wwwmautexasedu/cna/nmc6 October 16, 2011 Ward Cheney/David Kincaid

More information

MATH20602 Numerical Analysis 1

MATH20602 Numerical Analysis 1 M\cr NA Manchester Numerical Analysis MATH20602 Numerical Analysis 1 Martin Lotz School of Mathematics The University of Manchester Manchester, February 1, 2016 Outline General Course Information Introduction

More information

Ch01. Analysis of Algorithms

Ch01. Analysis of Algorithms Ch01. Analysis of Algorithms Input Algorithm Output Acknowledgement: Parts of slides in this presentation come from the materials accompanying the textbook Algorithm Design and Applications, by M. T. Goodrich

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 12 Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted for noncommercial,

More information

CSC236 Week 4. Larry Zhang

CSC236 Week 4. Larry Zhang CSC236 Week 4 Larry Zhang 1 Announcements PS2 due on Friday This week s tutorial: Exercises with big-oh PS1 feedback People generally did well Writing style need to be improved. This time the TAs are lenient,

More information

Lecture 1: Asymptotic Complexity. 1 These slides include material originally prepared by Dr.Ron Cytron, Dr. Jeremy Buhler, and Dr. Steve Cole.

Lecture 1: Asymptotic Complexity. 1 These slides include material originally prepared by Dr.Ron Cytron, Dr. Jeremy Buhler, and Dr. Steve Cole. Lecture 1: Asymptotic Complexity 1 These slides include material originally prepared by Dr.Ron Cytron, Dr. Jeremy Buhler, and Dr. Steve Cole. Announcements TA office hours officially start this week see

More information

Topic 17. Analysis of Algorithms

Topic 17. Analysis of Algorithms Topic 17 Analysis of Algorithms Analysis of Algorithms- Review Efficiency of an algorithm can be measured in terms of : Time complexity: a measure of the amount of time required to execute an algorithm

More information

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

6. Iterative Methods for Linear Systems. The stepwise approach to the solution... 6 Iterative Methods for Linear Systems The stepwise approach to the solution Miriam Mehl: 6 Iterative Methods for Linear Systems The stepwise approach to the solution, January 18, 2013 1 61 Large Sparse

More information

When we use asymptotic notation within an expression, the asymptotic notation is shorthand for an unspecified function satisfying the relation:

When we use asymptotic notation within an expression, the asymptotic notation is shorthand for an unspecified function satisfying the relation: CS 124 Section #1 Big-Oh, the Master Theorem, and MergeSort 1/29/2018 1 Big-Oh Notation 1.1 Definition Big-Oh notation is a way to describe the rate of growth of functions. In CS, we use it to describe

More information

Lecture 2. Fundamentals of the Analysis of Algorithm Efficiency

Lecture 2. Fundamentals of the Analysis of Algorithm Efficiency Lecture 2 Fundamentals of the Analysis of Algorithm Efficiency 1 Lecture Contents 1. Analysis Framework 2. Asymptotic Notations and Basic Efficiency Classes 3. Mathematical Analysis of Nonrecursive Algorithms

More information

1 Lecture 8: Interpolating polynomials.

1 Lecture 8: Interpolating polynomials. 1 Lecture 8: Interpolating polynomials. 1.1 Horner s method Before turning to the main idea of this part of the course, we consider how to evaluate a polynomial. Recall that a polynomial is an expression

More information

Fast Polynomial Multiplication

Fast Polynomial Multiplication Fast Polynomial Multiplication Marc Moreno Maza CS 9652, October 4, 2017 Plan Primitive roots of unity The discrete Fourier transform Convolution of polynomials The fast Fourier transform Fast convolution

More information

Elliptic Curves Spring 2013 Lecture #3 02/12/2013

Elliptic Curves Spring 2013 Lecture #3 02/12/2013 18.783 Elliptic Curves Spring 2013 Lecture #3 02/12/2013 3.1 Arithmetic in finite fields To make explicit computations with elliptic curves over finite fields, we need to know how to perform arithmetic

More information

Sequential Fast Fourier Transform

Sequential Fast Fourier Transform Sequential Fast Fourier Transform Departement Computerwetenschappen Room 03.018 (Celestijnenlaan 200A) albert-jan.yzelman@cs.kuleuven.be Includes material from slides by Prof. dr. Rob H. Bisseling Applications

More information

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane. Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 018 30 Lecture 30 April 7, 018 Fourier Series In this lecture,

More information

Fig. 1: Fourier series Examples

Fig. 1: Fourier series Examples FOURIER SERIES AND ITS SOME APPLICATIONS P. Sathyabama Assistant Professor, Department of Mathematics, Bharath collage of Science and Management, Thanjavur, Tamilnadu Abstract: The Fourier series, the

More information

Lecture 01 August 31, 2017

Lecture 01 August 31, 2017 Sketching Algorithms for Big Data Fall 2017 Prof. Jelani Nelson Lecture 01 August 31, 2017 Scribe: Vinh-Kha Le 1 Overview In this lecture, we overviewed the six main topics covered in the course, reviewed

More information

CSCE 222 Discrete Structures for Computing. Review for Exam 2. Dr. Hyunyoung Lee !!!

CSCE 222 Discrete Structures for Computing. Review for Exam 2. Dr. Hyunyoung Lee !!! CSCE 222 Discrete Structures for Computing Review for Exam 2 Dr. Hyunyoung Lee 1 Strategy for Exam Preparation - Start studying now (unless have already started) - Study class notes (lecture slides and

More information

Solving PDEs with Multigrid Methods p.1

Solving PDEs with Multigrid Methods p.1 Solving PDEs with Multigrid Methods Scott MacLachlan maclachl@colorado.edu Department of Applied Mathematics, University of Colorado at Boulder Solving PDEs with Multigrid Methods p.1 Support and Collaboration

More information

CSED233: Data Structures (2017F) Lecture4: Analysis of Algorithms

CSED233: Data Structures (2017F) Lecture4: Analysis of Algorithms (2017F) Lecture4: Analysis of Algorithms Daijin Kim CSE, POSTECH dkim@postech.ac.kr Running Time Most algorithms transform input objects into output objects. The running time of an algorithm typically

More information

CSE 241 Class 1. Jeremy Buhler. August 24,

CSE 241 Class 1. Jeremy Buhler. August 24, CSE 41 Class 1 Jeremy Buhler August 4, 015 Before class, write URL on board: http://classes.engineering.wustl.edu/cse41/. Also: Jeremy Buhler, Office: Jolley 506, 314-935-6180 1 Welcome and Introduction

More information

Introduction to Algorithms

Introduction to Algorithms Lecture 1 Introduction to Algorithms 1.1 Overview The purpose of this lecture is to give a brief overview of the topic of Algorithms and the kind of thinking it involves: why we focus on the subjects that

More information

Singular Value Decompsition

Singular Value Decompsition Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost

More information

University of the Virgin Islands, St. Thomas January 14, 2015 Algorithms and Programming for High Schoolers. Lecture 5

University of the Virgin Islands, St. Thomas January 14, 2015 Algorithms and Programming for High Schoolers. Lecture 5 University of the Virgin Islands, St. Thomas January 14, 2015 Algorithms and Programming for High Schoolers Numerical algorithms: Lecture 5 Today we ll cover algorithms for various numerical problems:

More information

Chapter 1 Divide and Conquer Polynomial Multiplication Algorithm Theory WS 2015/16 Fabian Kuhn

Chapter 1 Divide and Conquer Polynomial Multiplication Algorithm Theory WS 2015/16 Fabian Kuhn Chapter 1 Divide and Conquer Polynomial Multiplication Algorithm Theory WS 2015/16 Fabian Kuhn Formulation of the D&C principle Divide-and-conquer method for solving a problem instance of size n: 1. Divide

More information

Grade 11/12 Math Circles Fall Nov. 5 Recurrences, Part 2

Grade 11/12 Math Circles Fall Nov. 5 Recurrences, Part 2 1 Faculty of Mathematics Waterloo, Ontario Centre for Education in Mathematics and Computing Grade 11/12 Math Circles Fall 2014 - Nov. 5 Recurrences, Part 2 Running time of algorithms In computer science,

More information

Module 1: Analyzing the Efficiency of Algorithms

Module 1: Analyzing the Efficiency of Algorithms Module 1: Analyzing the Efficiency of Algorithms Dr. Natarajan Meghanathan Professor of Computer Science Jackson State University Jackson, MS 39217 E-mail: natarajan.meghanathan@jsums.edu What is an Algorithm?

More information

Discrete Mathematics CS October 17, 2006

Discrete Mathematics CS October 17, 2006 Discrete Mathematics CS 2610 October 17, 2006 Uncountable sets Theorem: The set of real numbers is uncountable. If a subset of a set is uncountable, then the set is uncountable. The cardinality of a subset

More information

Lecture 7 January 26, 2016

Lecture 7 January 26, 2016 MATH 262/CME 372: Applied Fourier Analysis and Winter 26 Elements of Modern Signal Processing Lecture 7 January 26, 26 Prof Emmanuel Candes Scribe: Carlos A Sing-Long, Edited by E Bates Outline Agenda:

More information

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.

More information

Linear Algebra and Eigenproblems

Linear Algebra and Eigenproblems Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details

More information

Algorithm efficiency can be measured in terms of: Time Space Other resources such as processors, network packets, etc.

Algorithm efficiency can be measured in terms of: Time Space Other resources such as processors, network packets, etc. Algorithms Analysis Algorithm efficiency can be measured in terms of: Time Space Other resources such as processors, network packets, etc. Algorithms analysis tends to focus on time: Techniques for measuring

More information

Lecture 2. More Algorithm Analysis, Math and MCSS By: Sarah Buchanan

Lecture 2. More Algorithm Analysis, Math and MCSS By: Sarah Buchanan Lecture 2 More Algorithm Analysis, Math and MCSS By: Sarah Buchanan Announcements Assignment #1 is posted online It is directly related to MCSS which we will be talking about today or Monday. There are

More information

Asymptotic Analysis 1

Asymptotic Analysis 1 Asymptotic Analysis 1 Last week, we discussed how to present algorithms using pseudocode. For example, we looked at an algorithm for singing the annoying song 99 Bottles of Beer on the Wall for arbitrary

More information

The O () notation. Definition: Let f(n), g(n) be functions of the natural (or real)

The O () notation. Definition: Let f(n), g(n) be functions of the natural (or real) The O () notation When analyzing the runtime of an algorithm, we want to consider the time required for large n. We also want to ignore constant factors (which often stem from tricks and do not indicate

More information

Defining Efficiency. 2: Analysis. Efficiency. Measuring efficiency. CSE 421: Intro Algorithms. Summer 2007 Larry Ruzzo

Defining Efficiency. 2: Analysis. Efficiency. Measuring efficiency. CSE 421: Intro Algorithms. Summer 2007 Larry Ruzzo CSE 421: Intro Algorithms 2: Analysis Summer 2007 Larry Ruzzo Defining Efficiency Runs fast on typical real problem instances Pro: sensible, bottom-line-oriented Con: moving target (diff computers, compilers,

More information