Bloom Features. Kwabena Boahen Bioengineering Department Stanford University Stanford CA, USA

Size: px
Start display at page:

Download "Bloom Features. Kwabena Boahen Bioengineering Department Stanford University Stanford CA, USA"

Transcription

1 2015 International Conference on Coputational Science and Coputational Intelligence Bloo Features Asok Cutkosky Coputer Science Departent Stanford University Stanford CA, USA Kwabena Boaen Bioengineering Departent Stanford University Stanford CA, USA Abstract We introduce a etod for function-fitting tat acieves ig accuracy wit a low eory footprint. For d-diensional data and any user-specified, we define a feature ap fro d to diensional uclidean space wit eory footprint O() tat scales as follows: As increases, te space of linear functions on our -diensional features approxiates any MAX (or boolean OR) function on te d-diensional inputs wit expected error inversely proportional to. Our etod is te only one in existence wit tis scaling tat can siultaneously run in O() tie, process real-value inputs, and approxiate non-linear functions, properties respectively not acieved by rando Fourier features, b-bit Minwise Hasing, and Vowpal Wabbit, tree copeting etods. We acieve all tree properties by using asing (O() space) to ipleent a sparse-atrix ultiply (O() tie) wit addition replaced by MAX (non-linear approxiation). As tese tecniques are inspired by te Bloo filter, we call te vectors produced by our apping Bloo features. We deonstrate tat te scaling pre-factors are reasonable by testing our etod on siulated (Diriclet distributions) and real (MNIST and webspa) datasets. Keywords-large-scale learning, learning teory I. LARNING WITH MMORY CONSTRAINTS Te use of randoness in order to save soe resource suc as eory (or coputation tie) at te potential expense of accuracy is an establised strategy trougout coputer science. A priary application of tese low-eory footprint etods is in large-scale settings in wic one ust process an extreely large aount of possibly ig-diensional data. In tis situation te original dataset ay not fit into eory and so one turns to etods tat reduce te eory footprint of te dataset but also allow te user to run soe type of analysis on te data. In acine learning, tese eory savings are acieved by etods tat use rando projections [1] to approxiate soe function f : R d R using a sall nuber of paraeters, but state of te art etods suffer fro particular drawbacks. Rando Fourier features [2] as poor asyptotic tie coplexity; Vowpal Wabbit (VW) [3] can only approxiate linear functions; and b-bit Minwise asing [4] can only operate on binary inputs. We present an algorit tat acieves a low eory footprint wile avoiding tese drawbacks. Siilar to te previous etods, for any we produce a apping φ : R d R suc tat as increases linear functions on R approxiate nonlinear functions in R d wit error inversely proportional to. However our eory footprint increases linearly wit, and so represents a trade-off between eory and accuracy in function fitting. We call te vectors in R produced by our apping Bloo features (Definition A.1), as our etod was inspired by te Bloo filter data structure [5]. To copute Bloo features, we coose k as functions to assign eac coponent of a vector x in R d to k coponents of φ(x) in R. We copute te it coponent of φ(x) by taking te MAX of all te coponents of x tat are apped to tis it coponent (Figure 1). Tis is analogous to ultiplication by a sparse d atrix (specified by te k as functions) wit addition replaced by MAX. Using as functions and sparsity allows us to ave low eory footprint and sall tie coplexity wile te substitution of MAX for addition allows us to copute non-linear functions on real-valued inputs. x: φ(x): Fig. 1. ac coordinate of x is apped to k rando coordinates in φ(x) by as functions. Te final value of eac coordinate of φ(x) is given by taking te axiu of te values ased to it. In tis case k =2and = /15 $ I DOI /CSCI

2 In Section II, we introduce Bloo features and describe teir properties. In Section III, we epirically validate Bloo features on te freely available MNIST [6], and webspa [7] datasets. We also provide results on siulated data. We state foral results and proofs in an appendix. II. BLOOM FATURS Approxiating functions wit Bloo features offers a good tradeoff between space, tie and accuracy. Given input/output pairs (x i,y i ) R d R, tis proble ay be forulated as follows: Find f w : R d R of te for f w (x) = φ(x),w were a, b indicates inner-product, suc tat f w (x i ) y i, were φ : R d R is a fixed ap and w R is cosen to acieve te best approxiation. Te degree of approxiation is easured by soe loss function L. For exaple, L(f w,x,y) = (f w (x) y) 2 is a coon coice. Linear regression can be forulated in tis anner wit φ(x) =x and quadratic-kernel support vector regression corresponds to φ(x) equal to a vector of all 2 nd order interactions in x. Forulating function-approxiation as a linearap on a set of -diensional features φ(x), provided by a fixed apping φ, introduces a tradeoff between resources and accuracy. As increases, φ aps into iger diensional spaces, resulting in ore accurate function-fitting troug inner products wit φ(x), but requiring ore tie and space to be coputed. A desirable φ will obtain a good tradeoff between space, tie and accuracy. Te Bloo feature ap acieves a good tradeoff by using as functions to specify a sparse atrix and replacing te addition in atrix ultiplication by MAX. Tis ap φ operates on d-diensional non-negative inputs tat is, eleents of R d 0. It can approxiate any linear cobination of binary OR functions or real-valued MAX functions wit error tat decreases inversely proportional to (Teores A.4, A.8) and can be coputed in O() space and O() tie (Proposition A.2). Wile its space scaling is te sae as tree state-of-te-art etods, its tie scaling is n to n/ ties better (Table I). TABL I COMPARISON BTWN MTHODS Feature Tie Non-Binary Non-linear Bloo O() Yes Yes Fourier O(n) Yes Yes b-bit Minwise O( n 2b ) No Yes VW O(n) Yes No III. MPIRICAL RSULTS To verify tat Bloo features resource scaling factors are reasonable and confir tat its teoretical perforance generalizes to te real-world, we tested it on siulated data (binary and real-valued) and real data (MNIST and webspa). A. Siulated Data We first tested Bloo features on a siulated classification task wit binary vectors. Te binary vectors were drawn fro {0, 1} 100 wit 10 randoly selected non-zero entries. Te two classes were defined by an OR function coputed on eiter 3, 5 or 10 randoly cosen bits. We cose a subset of te data tat ad an equal fraction belonging to eac class. We used Bloo features wit = 100, 1000, and k = 10 log(2), te optial value for a Bloo filter. We found tat Bloo features acieve low MS using values significantly saller tan ( ) 100 F, te nuber of possible OR functions of fan-in F (Figure 2). Mean Squared rror value Fan-in: 10 Fan-in: 5 Fan-in: 3 Fig. 2. Mean squared error vs for a boolean OR of fan-in 3 and 10, wit d = 100. Horizontal lines indicate best linear regression error for coparison. rror bars represent one standard deviation. Next, we tested Bloo features on a siulated realvalued classification task consisting of two classes defined by two distinct, 100-diensional Diriclet distributions wit paraeters α and 3α+β 4 respectively. α and β were drawn uniforly at rando fro (0, 1) 100. Tis particular coice results in a Bayes risk of 0.3% (error of optial classifier). We copared Bloo features to rando Fourier features wit ranging fro 50 to for bot. k was set to 10 log(2) for Bloo features; b-bit Minwise asing was not included because tis etod only

3 applies to binary inputs. We found tat Fourier features were ore accurate for sall wereas Bloo features were ore accurate for large (Figure 3). Accuracy Bloo Features Fourier Features Fig. 3. Bloo features vs. rando Fourier Features on siulated dataset. rror bars represent one standard deviation. We ten proceeded to test Bloo features on ore realistic data by classifying te MNIST [6] and webspa [7] datasets. For all tasks, we applied one-vs-all linear classification on a Bloo feature representation of te data. B. MNIST Te MNIST dataset contains grayscale iages of andwritten digits 0 troug 9, wic are split into a training and test set of size and respectively. Te task is to deterine wic digit a given iage represents. We consider te perutation-invariant for of te task, in wic one cannot take advantage of any previously known structure in te data (suc as te fact tat te inputs are 2D iages). We cose k via cross-validation on = 400, wic gave k = 4. We terefore used k = /100 (rounded to te nearest integer) for all oter values of. We copared Bloo features perforance on MNIST wit tat of rando Fourier Features and b- bit Minwise asing, wic ave te sae eory footprint. We did not include VW because it can only classify linearly separable data; it is well-known tat MNIST is not linearly separable. In te case of Minwise asing, we converted te iages to binary vectors by tresolding values to 1, an easy and natural process for MNIST. For oter datasets (e.g. our siulation in Figure 3), tis ay not be true. We found tat te perforance gap between Bloo and Fourier features narrowed as increased (Table II), as expected fro our results wit siulated data (Figure 3). And bot out-perfored Minwise asing, deonstrating te advantage of using realvalued vectors. We also copared Bloo features perforance on MNIST wit several state-of-te-art algorits tat ave a uc larger eory footprint as easured by paraeter count (taken fro Table 1 of [8]). One of tese algorits (Maxout MLP, described in [8]) is siilar in spirit to Bloo features. In tis etod, one trains a ulti-layer perceptron wose activation function takes te axiu of its inputs. We found tat Bloo features use 95% fewer paraeters tan any of tese algorits wile acieving an error-rate witin a factor of two of te lowest result (1.6% versus 0.79%). TABL II MNIST PRFORMANC (AVRAG RROR) Bloo Fourier b-bit Minwise % 5.3% 11.1% % 3.1% 7.4% % 1.6% 2.8% TABL III MNIST PRFORMANC (OTHR MTHODS) Metod rror Paraeter Count ReLU MLP [9] 1.05% Maxout MLP [8] 0.94% Manifold Tangent Classifier [10] 0.81% DBM (wit dropout) [11] 0.79% Bloo Features 1.6% C. webspa Te webspa [7] dataset consists of sparse vectors of trigra counts representing eiter spa or non-spa docuents. Te data ave a diensionality of , of wic only coponents are ever nonzero. webspa is nearly linearly separable, wit linear classification accuracies in excess of 99%. Tus, we consider classifying webspa as a deonstration tat Bloo features are still able to capture linear relationsips. To classify webspa we selected 80% of te data to be a training set and 20% to be a testing set. We copared Bloo features perforance on webspa wit tat of rando Fourier Features, b-bit Minwise asing, and VW for = 100, 1000, All tree coices use uc less

4 eory tan is needed to represent te original features (up to vs ). webspa s sparse binary features and linearly separability are exactly te conditions required by b-bit Minwise asing and VW, respectively, ence tey perfored te best (Table IV). Neverteless, across te range of values tested, Bloo Features error was no worse tan ties tat of tese etods. TABL IV WBSPAM PRFORMANC (AVRAG RROR) Bloo Fourier b-bit Minwise VW % 26.4% 10.8% 8.1% % 14.7% 2.1% 1.9% % 7.9% 0.8% 1.1% IV. CONCLUSIONS We ave introduced Bloo features and analyzed teir perforance. Bloo feature representations are eory-efficient (O()) and can be coputed very quickly (O()). As teir diensionality increases, tey approxiate boolean functions and non-linear real-valued functions wit error decreasing inversely proportionally to. We prove tese results in te Appendix and sow tat Bloo features operate by approxiating a ig-diensional RKHS consisting of all interactions aong inputs. We deonstrated te practical viability of Bloo features using siulated data and te MNIST and webspa datasets. On te siulated and MNIST datasets, Bloo features represented non-linear functions on real-valued inputs accurately and efficiently. On te webspa dataset, Bloo features were copetitive wit ore restrictive etods suc as b-bit Minwise asing and VW wen representing linear functions on binary inputs. Tus Bloo features not only possess teoretical advantages in ters of eiter andling real-valued inputs, coputation tie, or non-linear approxiation over oter etods, tey also copare favorably on practical tasks in ters of test error for a given eory footprint. RFRNCS [1] Willia B Jonson and Jora Lindenstrauss. xtensions of lipscitz appings into a ilbert space. Conteporary ateatics, 26( ):1, [2] Ali Raii and Benjain Rect. Rando Features for Large Scale Kernel Macines. In Advances in Neural Inforation Processing Systes, [3] W Kilian, D Anirban, L Jon, S Alex, and A Jos. Feature Hasing for Large Scale Multitask Learning Feature Hasing for Large Scale Multitask Learning. In International Conference on Macine Learning (ICML), pages , [4] Ping Li and Cristian König. b-bit inwise asing. In Proceedings of te 19t international conference on World wide web, pages ACM, [5] B Bloo. Space/Tie trade-offs in as coding wit allowable errors. Counications of te ACM, 13(7): , [6] L Yann, B Léon, B Yosua, and H Patrick. Gradient-Based Learning Applied to Docunet Recognition. Proceedings of te I, 86(11): , [7] Steve Webb, Jaes Caverlee, and Calton Pu. Introducing te webb spa corpus: Using eail spa to identify web spa autoatically. In CAS, [8] G I J, W David, M Medi, C Aaron, and B Yosua. Maxout Networks. Journal of Macine Learning Researc (JMLR), 28(3): , [9] Nitis Srivastava. Iproving neural networks wit dropout. PD tesis, University of Toronto, [10] Sala Rifai, Yann N Daupin, Pascal Vincent, Yosua Bengio, and Xavier Muller. Te anifold tangent classifier. In Advances in Neural Inforation Processing Systes, pages , [11] Geoffrey Hinton, Nitis Srivastava, Alex Krizevsky, Ilya Sutskever, and Ruslan R Salakutdinov. Iproving neural networks by preventing co-adaptation of feature detectors. arxiv preprint arxiv: , [12] Micael Mitzenacer and li Upfal. Probability and coputing: Randoized algorits and probabilistic analysis. Cabridge University Press, APPNDIX Definition A.1. Bloo Features Suppose 1,..., k are functions apping N to {0,..., 1} drawn at rando fro a pair-wise independent faily of as functions. For x R d 0, define φ (x) R 0 by φ (x) i = ax j,l: l (j)=i x j were te subscript in φ is intended to epasize tat φ (x) depends on te coice of as functions. Proposition A.2. Coputation Suppose x R d wit n non-zero coponents, and let φ (x) be a Bloo feature ap using k as functions. Ten φ (x) can be coputed in O(kn) tie wen vectors are represented in sparse forat (only non-zero quantities stored), and in O(d + kn + ) tie wen vectors are represented in dense forat. Proposition A.3. Meory Standard Bloo filter analysis [12] suggests a value for k: If our inputs x ave n non-zero coponents, ten we sould set k = log(2) n. Using tis value of k and Proposition A.2, we get a coplexity of O(kn) =O(). Teore A.4. Binary Input Approxiation Suppose B(x) is a boolean function of fan-in F and

5 φ (x) : R d R is a Bloo feature ap wit k as functions (see Definition A.1). Let n be te nuber of set bits of soe vector x {0, 1} d. Let p =1 (1 1 )k be te probability tat a bit of x is ased to a given coponent of φ (x). Ten for sufficiently large tere exists w R suc tat [φ (x) w ]=B(x) ( [(φ (x) w B(x)) 2 ]=O 2 F p F (1 1/) kn Proof: Our proof as two steps. First we observe tat we can write B(x) as a linear cobination of ORs involving no negations: B(x) = s i a i(x i1 x iz ) wit a i R, z F and s 2 F. Ten we sow tat x i1 x iz φ (x) w,i for soe appropriate w,i (see Lea A.5). Substituting tis approxiation in to te expression for B(x) copletes te proof wit w = i a iw,i. To prove te first step, we consider OR functions as vectors in R 2F (defined by teir trut tables), and sow tat tey are linearly independent in tis space. To see tis, first set y j =1 x j. Ten by De Morgan s rules we ave x i1 x iz = 1 y i1 y iz. Tus tere is a 1-1 linear ap fro OR functions to onoials y i1 y iz. Now since te set of distinct onoials is linearly independent, so are te OR functions. Lea A.5. Suppose C 1 (x) and C 2 (x) are OR functions of fan-in F 1 and F 2. Furter, define F as te nuber of inputs sared by C 1 and C 2 and let F = (F 1 + F 2 ) F. Suppose φ (x) is a Bloo feature ap wit k as functions. Ten for sufficiently large tere exists w 1 and w2 suc tat: [φ (x) w i ]=C i (x) Var(φ (x) w i ) 4 p F i (1 p) n Cov(φ (x) w 1,φ (x) w 2 ) 2 p F (1 p) n were p,, n, x were defined previously (see Teore A.4). Proof: If φ (x) s coponents were bot zeroean and independent, ten we can prove tis result wit bounds iproved by a factor of 4 (Lea A.7). To prove tis result for te non-zero-ean, non-independent φ, we subtract φ (x) s first /2 coponents fro its second /2 coponents to for a new vector φ (x) tat is alf te size but is now zero-ean wile still aving non-independent ) coponents. Since φ (x) is alf te size, we lose a factor of 2 in te bounds. Now we coplete te proof by sowing tat te lack of independence between coponents introduces anoter factor of 2 in te bounds. Te factor of 2 loosening due to te lack of independence coes fro distributing expectations over products. Specifically, we need to sow tat distributing te expectation [φ (x) i φ (x) j ] over te product accrues a error tat goes down as increases. To do tis we sow tat φ (x) i concentrates about its ean. Let e be te nuber of zero bits in φ (x) and q = e. Mitzenacer and Upfal sow [12] for Bloo filters (wic applies to Bloo features in te binary case) tat P( q [q] > λ ) < 2exp( 2λ 2 /). Tus [q 2 ] (1 2exp( 2λ 2 /))([q] λ/) 2 [q 2 ] (1 2exp( 2λ 2 /))([q]+λ/) 2 +2exp( 2λ 2 /) So tat for any δ, for sufficiently large, we ust ave 1 1 δ [q] 2 [q 2 ] (1 + δ) [q] 2.Now [φ (x) i φ (x) j ]= t P(q = t)(1 t)(1 t 1 ) = [q]+ 1 [q 2 ] Tus for any δ, for sufficiently large, [φ (x) i φ (x) j ] is witin a factor of 1 + δ of (1 [q]) 2 = [φ (x) i ] [φ (x) j ]. Coosing δ =1, for sufficiently large te error fro te independence assuption is bounded by a factor of 2. Definition A.6. Independent Features Let p (0, 1). Suppose z + i,j,z i,j are independent Bernoulli rando variables wit te sae ean p for i {0,..., 1}, j {0,...,d 1}. Ten define T + (x) i = ax x j j:z + i,j =1 T (x) i = ax x j j:z i,j =1 φ z (x) i = T + (x) i T (x) i were ere te subscript z indicates tat φ z depends on te values of te variables z + i,j and z i,j. Lea A.7. Let C 1, C 2, F 1, F 2, F, F, x and n be as defined previously (Teore A.4). Let φ z (x) be an

6 independent feature wit paraeter p (see Definition A.6). Ten tere exists w 1 z,w 2 z R suc tat: z [φ z (x) w i z]=c i (x) Var(φ z (x) w i z) 1 p F i (1 p) n Cov(φ z (x) w 1 z,φ z (x) w 2 z) 1 2 p F (1 p) n Proof: We ll prove te bias and variance results for C = C 1. Te results for C 2 are syetric. Let C = x c1 x cf. We start by defining Q n =1 (1 p) n, Z + j = F i=1 z+ c i,j, and Z j = F i=1 z c i,j. Te following facts will be useful: [Z + j ]= [Z j ]=pf (1) [Z + j T + (x) j ]=p F C(x)+p F (1 C(x))Q n (2) [Z + j T (x) j ]=p F Q n (3) Wit (r z ) j = Z + j Z j, soe algebra gives: [φ z (x) r z ]=[φ z (x) 1 (r z ) 1 ] z z =2 z [T + 1 Z+ 1 ] 2 z [T 1 Z+ 1 ] =2p F (1 Q n )C(x) For te variance, we copute Var(φ z (x) r z )=Var(φ z (x) 1 (r z ) 1 ) [(φ z (x) 1 (r z ) 1 ) 2 ]=[(Z + Z ) 2 (T 1 + T 1 )2 ] fro wic we obtain (fro equations 1, 2, 3): Var(φ z (x) r z ) 4p F (1 Q n ) Nowifweset(w z ) j = 1 2 p F (1 Q n ) 1 1 (r z) j ten we recover te expectation and variance stateents. Te covariance stateent is coputed siilarly. Teore A.8. Approxiating a space of MAX functions Suppose φ (x) is a Bloo feature constructed wit k as functions. Suppose K(x, y) is a scaled inner product of MAX functions as defined below. Ten: ( P [ 1 φ (x) 1 φ (y)] = K(x, y) ) φ sup (x) φ(y) x,y D n K(x, y) >ɛ as long as ( > 2 ɛ (log(1/δ)+2log 2 ( ) )) d 2 n <δ Were D n [0, 1] d is te set of vectors wit at ost n non-zero coponents. K(x, y) is given by te following construction. For x [0, 1] d, let MAX q (x) R (d q) be te vector obtained by applying all q-ary MAX functions to te entries of x. Define K(x, y) =ψ(x) ψ(y) wit ψ(x) = d q=1 (1 p) d q p q MAX q (x) were p =1 (1 1 )k. Tus K(x, y) is te kernel of te RKHS H defined by H d d ) = R( q = R 2d q=1 wit v(x) =v ψ(x) for v H. Proof: First we copute te expectation: [ 1 φ (x) 1 φ (y)] = [φ (x) 1 φ (y) 1 ] Let A q be te event tat tere are exactly q values of t suc tat l (t) =1for soe l. Ten P (A q )= (1 p) d q p q( d q). Furter, [φ (x) 1 φ (y) 1 A q ]= ( d q) 1MAXq (x) MAX q (y) n q=1 [ φ (x) φ(y) ] = P (A q ) ( d q) 1MAXq (x) MAX q (y) =K(x, y) For te concentration inequality, we apply te Azua-Hoeffding inequality to te Doob artingale given by B i = 1 [φ (x) φ (y) φ (x) 1,...,φ (y) i ] Since eac coponent of φ (x) is in [0, 1], B i+1 B i 1, and so ( ) φ P (x) φ(y) K(x, y) >ɛ 2exp( ɛ2 2 ) Te axiu error occurs at corners of D n,soby applying a union bound over all ( d 2 n) pairs x, y we bound te probability of error ore tan ɛ by 2 ( d n) 2 exp( ɛ 2 /2) = δ Solve for to prove te teore

1 Proving the Fundamental Theorem of Statistical Learning

1 Proving the Fundamental Theorem of Statistical Learning THEORETICAL MACHINE LEARNING COS 5 LECTURE #7 APRIL 5, 6 LECTURER: ELAD HAZAN NAME: FERMI MA ANDDANIEL SUO oving te Fundaental Teore of Statistical Learning In tis section, we prove te following: Teore.

More information

lecture 35: Linear Multistep Mehods: Truncation Error

lecture 35: Linear Multistep Mehods: Truncation Error 88 lecture 5: Linear Multistep Meods: Truncation Error 5.5 Linear ultistep etods One-step etods construct an approxiate solution x k+ x(t k+ ) using only one previous approxiation, x k. Tis approac enoys

More information

Neural Networks Trained with the EEM Algorithm: Tuning the Smoothing Parameter

Neural Networks Trained with the EEM Algorithm: Tuning the Smoothing Parameter eural etworks Trained wit te EEM Algorit: Tuning te Sooting Paraeter JORGE M. SATOS,2, JOAQUIM MARQUES DE SÁ AD LUÍS A. ALEXADRE 3 Intituto de Engenaria Bioédica, Porto, Portugal 2 Instituto Superior de

More information

Derivative at a point

Derivative at a point Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Derivative at a point Wat you need to know already: Te concept of liit and basic etods for coputing liits. Wat you can

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

Supplementary Materials: Proofs and Technical Details for Parsimonious Tensor Response Regression Lexin Li and Xin Zhang

Supplementary Materials: Proofs and Technical Details for Parsimonious Tensor Response Regression Lexin Li and Xin Zhang Suppleentary Materials: Proofs and Tecnical Details for Parsionious Tensor Response Regression Lexin Li and Xin Zang A Soe preliinary results We will apply te following two results repeatedly. For a positive

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Estimation for the Parameters of the Exponentiated Exponential Distribution Using a Median Ranked Set Sampling

Estimation for the Parameters of the Exponentiated Exponential Distribution Using a Median Ranked Set Sampling Journal of Modern Applied Statistical Metods Volue 14 Issue 1 Article 19 5-1-015 Estiation for te Paraeters of te Exponentiated Exponential Distribution Using a Median Ranked Set Sapling Monjed H. Sau

More information

Stationary Gaussian Markov processes as limits of stationary autoregressive time series

Stationary Gaussian Markov processes as limits of stationary autoregressive time series Stationary Gaussian Markov processes as liits of stationary autoregressive tie series Pilip A. rnst 1,, Lawrence D. Brown 2,, Larry Sepp 3,, Robert L. Wolpert 4, Abstract We consider te class, C p, of

More information

Estimating the Density of a Conditional Expectation

Estimating the Density of a Conditional Expectation Estiating te Density of a Conditional Expectation Sauel G. Steckley Sane G. Henderson David Ruppert Ran Yang Daniel W. Apley Jerey Stau Abstract In tis paper, we analyze etods for estiating te density

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Regularized Regression

Regularized Regression Regularized Regression David M. Blei Columbia University December 5, 205 Modern regression problems are ig dimensional, wic means tat te number of covariates p is large. In practice statisticians regularize

More information

Combining Classifiers

Combining Classifiers Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

5.1 The derivative or the gradient of a curve. Definition and finding the gradient from first principles

5.1 The derivative or the gradient of a curve. Definition and finding the gradient from first principles Capter 5: Dierentiation In tis capter, we will study: 51 e derivative or te gradient o a curve Deinition and inding te gradient ro irst principles 5 Forulas or derivatives 5 e equation o te tangent line

More information

LAB #3: ELECTROSTATIC FIELD COMPUTATION

LAB #3: ELECTROSTATIC FIELD COMPUTATION ECE 306 Revised: 1-6-00 LAB #3: ELECTROSTATIC FIELD COMPUTATION Purpose During tis lab you will investigate te ways in wic te electrostatic field can be teoretically predicted. Bot analytic and nuerical

More information

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

A Theoretical Analysis of a Warm Start Technique

A Theoretical Analysis of a Warm Start Technique A Theoretical Analysis of a War Start Technique Martin A. Zinkevich Yahoo! Labs 701 First Avenue Sunnyvale, CA Abstract Batch gradient descent looks at every data point for every step, which is wasteful

More information

Determining Limits of Thermal NDT of Thick Graphite/Epoxy Composites

Determining Limits of Thermal NDT of Thick Graphite/Epoxy Composites ECNDT 006 - We.3.8.1 Deterining Liits of Teral NDT of Tick Grapite/Epoy Coposites Vladiir VAVILOV Institute of Introscopy Tosk Russia Abstract. Te known approac to inspecting tin coposites by using infrared

More information

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t. CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

Support Vector Machines. Goals for the lecture

Support Vector Machines. Goals for the lecture Support Vector Machines Mark Craven and David Page Coputer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Soe of the slides in these lectures have been adapted/borrowed fro aterials developed

More information

A KERNEL APPROACH TO ESTIMATING THE DENSITY OF A CONDITIONAL EXPECTATION. Samuel G. Steckley Shane G. Henderson

A KERNEL APPROACH TO ESTIMATING THE DENSITY OF A CONDITIONAL EXPECTATION. Samuel G. Steckley Shane G. Henderson Proceedings of te 3 Winter Siulation Conference S Cick P J Sáncez D Ferrin and D J Morrice eds A KERNEL APPROACH TO ESTIMATING THE DENSITY OF A CONDITIONAL EXPECTATION Sauel G Steckley Sane G Henderson

More information

Numerical Solution for Non-Stationary Heat Equation in Cooling of Computer Radiator System

Numerical Solution for Non-Stationary Heat Equation in Cooling of Computer Radiator System (JZS) Journal of Zankoy Sulaiani, 9, 1(1) Part A (97-1) A119 Nuerical Solution for Non-Stationary Heat Equation in Cooling of Coputer Radiator Syste Aree A. Maad*, Faraidun K. Haa Sal**, and Najadin W.

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

Submanifold density estimation

Submanifold density estimation Subanifold density estiation Arkadas Ozakin Georgia Tec Researc Institute Georgia Insitute of Tecnology arkadas.ozakin@gtri.gatec.edu Alexander Gray College of Coputing Georgia Institute of Tecnology agray@cc.gatec.edu

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

Polygonal Designs: Existence and Construction

Polygonal Designs: Existence and Construction Polygonal Designs: Existence and Construction John Hegean Departent of Matheatics, Stanford University, Stanford, CA 9405 Jeff Langford Departent of Matheatics, Drake University, Des Moines, IA 5011 G

More information

PAC-Bayes Analysis Of Maximum Entropy Learning

PAC-Bayes Analysis Of Maximum Entropy Learning PAC-Bayes Analysis Of Maxiu Entropy Learning John Shawe-Taylor and David R. Hardoon Centre for Coputational Statistics and Machine Learning Departent of Coputer Science University College London, UK, WC1E

More information

1 Proof of learning bounds

1 Proof of learning bounds COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #4 Scribe: Akshay Mittal February 13, 2013 1 Proof of learning bounds For intuition of the following theore, suppose there exists a

More information

A KERNEL APPROACH TO ESTIMATING THE DENSITY OF A CONDITIONAL EXPECTATION. Samuel G. Steckley Shane G. Henderson

A KERNEL APPROACH TO ESTIMATING THE DENSITY OF A CONDITIONAL EXPECTATION. Samuel G. Steckley Shane G. Henderson Proceedings of te 3 Winter Siulation Conference S Cick P J Sáncez D Ferrin and D J Morrice eds A KERNEL APPROACH TO ESTIMATING THE DENSITY OF A CONDITIONAL EXPECTATION Sauel G Steckley Sane G Henderson

More information

Uniform Approximation and Bernstein Polynomials with Coefficients in the Unit Interval

Uniform Approximation and Bernstein Polynomials with Coefficients in the Unit Interval Unifor Approxiation and Bernstein Polynoials with Coefficients in the Unit Interval Weiang Qian and Marc D. Riedel Electrical and Coputer Engineering, University of Minnesota 200 Union St. S.E. Minneapolis,

More information

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words)

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words) 1 A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine (1900 words) Contact: Jerry Farlow Dept of Matheatics Univeristy of Maine Orono, ME 04469 Tel (07) 866-3540 Eail: farlow@ath.uaine.edu

More information

Support Vector Machines. Maximizing the Margin

Support Vector Machines. Maximizing the Margin Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

1 Bounding the Margin

1 Bounding the Margin COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost

More information

A remark on a success rate model for DPA and CPA

A remark on a success rate model for DPA and CPA A reark on a success rate odel for DPA and CPA A. Wieers, BSI Version 0.5 andreas.wieers@bsi.bund.de Septeber 5, 2018 Abstract The success rate is the ost coon evaluation etric for easuring the perforance

More information

Exploring the Exponential Integrators with Krylov Subspace Algorithms for Nonlinear Circuit Simulation

Exploring the Exponential Integrators with Krylov Subspace Algorithms for Nonlinear Circuit Simulation Exploring te Exponential Integrators wit Krylov Subspace Algorits for Nonlinear Circuit Siulation Xinyuan Wang, Hao Zuang +, Cung-Kuan Ceng CSE and ECE Departents, UC San Diego, La Jolla, CA, USA + ANSYS,

More information

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay

More information

A Low-Complexity Congestion Control and Scheduling Algorithm for Multihop Wireless Networks with Order-Optimal Per-Flow Delay

A Low-Complexity Congestion Control and Scheduling Algorithm for Multihop Wireless Networks with Order-Optimal Per-Flow Delay A Low-Coplexity Congestion Control and Scheduling Algorith for Multihop Wireless Networks with Order-Optial Per-Flow Delay Po-Kai Huang, Xiaojun Lin, and Chih-Chun Wang School of Electrical and Coputer

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

c hc h c h. Chapter Since E n L 2 in Eq. 39-4, we see that if L is doubled, then E 1 becomes (2.6 ev)(2) 2 = 0.65 ev.

c hc h c h. Chapter Since E n L 2 in Eq. 39-4, we see that if L is doubled, then E 1 becomes (2.6 ev)(2) 2 = 0.65 ev. Capter 39 Since n L in q 39-4, we see tat if L is doubled, ten becoes (6 ev)() = 065 ev We first note tat since = 666 0 34 J s and c = 998 0 8 /s, 34 8 c6 66 0 J sc 998 0 / s c 40eV n 9 9 60 0 J / ev 0

More information

The research of the rst author was supported in part by an Information Technology

The research of the rst author was supported in part by an Information Technology Tecnical Report 95-376 Absorbing Boundary Conditions for te Scrodinger Equation Toas Fevens Hong Jiang February 16, 1995 Te researc of te rst autor was supported in part by an Inforation Tecnology Researc

More information

Numerical Differentiation

Numerical Differentiation Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function

More information

. The univariate situation. It is well-known for a long tie that denoinators of Pade approxiants can be considered as orthogonal polynoials with respe

. The univariate situation. It is well-known for a long tie that denoinators of Pade approxiants can be considered as orthogonal polynoials with respe PROPERTIES OF MULTIVARIATE HOMOGENEOUS ORTHOGONAL POLYNOMIALS Brahi Benouahane y Annie Cuyt? Keywords Abstract It is well-known that the denoinators of Pade approxiants can be considered as orthogonal

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a

More information

Design of Spatially Coupled LDPC Codes over GF(q) for Windowed Decoding

Design of Spatially Coupled LDPC Codes over GF(q) for Windowed Decoding IEEE TRANSACTIONS ON INFORMATION THEORY (SUBMITTED PAPER) 1 Design of Spatially Coupled LDPC Codes over GF(q) for Windowed Decoding Lai Wei, Student Meber, IEEE, David G. M. Mitchell, Meber, IEEE, Thoas

More information

Testing equality of variances for multiple univariate normal populations

Testing equality of variances for multiple univariate normal populations University of Wollongong Research Online Centre for Statistical & Survey Methodology Working Paper Series Faculty of Engineering and Inforation Sciences 0 esting equality of variances for ultiple univariate

More information

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY (Section 3.2: Derivative Functions and Differentiability) 3.2.1 SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY LEARNING OBJECTIVES Know, understand, and apply te Limit Definition of te Derivative

More information

In this chapter, we consider several graph-theoretic and probabilistic models

In this chapter, we consider several graph-theoretic and probabilistic models THREE ONE GRAPH-THEORETIC AND STATISTICAL MODELS 3.1 INTRODUCTION In this chapter, we consider several graph-theoretic and probabilistic odels for a social network, which we do under different assuptions

More information

Soft-margin SVM can address linearly separable problems with outliers

Soft-margin SVM can address linearly separable problems with outliers Non-linear Support Vector Machines Non-linearly separable probles Hard-argin SVM can address linearly separable probles Soft-argin SVM can address linearly separable probles with outliers Non-linearly

More information

The Transactional Nature of Quantum Information

The Transactional Nature of Quantum Information The Transactional Nature of Quantu Inforation Subhash Kak Departent of Coputer Science Oklahoa State University Stillwater, OK 7478 ABSTRACT Inforation, in its counications sense, is a transactional property.

More information

INFORMATION THEORETIC STRUCTURE LEARNING WITH CONFIDENCE

INFORMATION THEORETIC STRUCTURE LEARNING WITH CONFIDENCE INFORMATION THEORETIC STRUCTURE LEARNING WITH CONFIDENCE Kevin R. Moon Yale University Departent of Genetics New Haven, Connecticut, U.S.A Morteza Nosad, Salie Yasaei Seke, Alfred O. Hero III University

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

Finite fields. and we ve used it in various examples and homework problems. In these notes I will introduce more finite fields

Finite fields. and we ve used it in various examples and homework problems. In these notes I will introduce more finite fields Finite fields I talked in class about the field with two eleents F 2 = {, } and we ve used it in various eaples and hoework probles. In these notes I will introduce ore finite fields F p = {,,...,p } for

More information

Interactive Markov Models of Evolutionary Algorithms

Interactive Markov Models of Evolutionary Algorithms Cleveland State University EngagedScholarship@CSU Electrical Engineering & Coputer Science Faculty Publications Electrical Engineering & Coputer Science Departent 2015 Interactive Markov Models of Evolutionary

More information

Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy

Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy Moammad Ali Keyvanrad a, Moammad Medi Homayounpour a a Laboratory for Intelligent Multimedia Processing (LIMP), Computer

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

Bootstrapping Dependent Data

Bootstrapping Dependent Data Bootstrapping Dependent Data One of the key issues confronting bootstrap resapling approxiations is how to deal with dependent data. Consider a sequence fx t g n t= of dependent rando variables. Clearly

More information

Physics 215 Winter The Density Matrix

Physics 215 Winter The Density Matrix Physics 215 Winter 2018 The Density Matrix The quantu space of states is a Hilbert space H. Any state vector ψ H is a pure state. Since any linear cobination of eleents of H are also an eleent of H, it

More information

The proofs of Theorem 1-3 are along the lines of Wied and Galeano (2013).

The proofs of Theorem 1-3 are along the lines of Wied and Galeano (2013). A Appendix: Proofs The proofs of Theore 1-3 are along the lines of Wied and Galeano (2013) Proof of Theore 1 Let D[d 1, d 2 ] be the space of càdlàg functions on the interval [d 1, d 2 ] equipped with

More information

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households Volume 29, Issue 3 Existence of competitive equilibrium in economies wit multi-member ouseolds Noriisa Sato Graduate Scool of Economics, Waseda University Abstract Tis paper focuses on te existence of

More information

Fairness via priority scheduling

Fairness via priority scheduling Fairness via priority scheduling Veeraruna Kavitha, N Heachandra and Debayan Das IEOR, IIT Bobay, Mubai, 400076, India vavitha,nh,debayan}@iitbacin Abstract In the context of ulti-agent resource allocation

More information

arxiv: v1 [cs.ds] 3 Feb 2014

arxiv: v1 [cs.ds] 3 Feb 2014 arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/

More information

Probability Distributions

Probability Distributions Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples

More information

Chapter 5 FINITE DIFFERENCE METHOD (FDM)

Chapter 5 FINITE DIFFERENCE METHOD (FDM) MEE7 Computer Modeling Tecniques in Engineering Capter 5 FINITE DIFFERENCE METHOD (FDM) 5. Introduction to FDM Te finite difference tecniques are based upon approximations wic permit replacing differential

More information

1 Generalization bounds based on Rademacher complexity

1 Generalization bounds based on Rademacher complexity COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #0 Scribe: Suqi Liu March 07, 08 Last tie we started proving this very general result about how quickly the epirical average converges

More information

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these. Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra

More information

1 Introduction to Optimization

1 Introduction to Optimization Unconstrained Convex Optimization 2 1 Introduction to Optimization Given a general optimization problem of te form min x f(x) (1.1) were f : R n R. Sometimes te problem as constraints (we are only interested

More information

A Smoothed Boosting Algorithm Using Probabilistic Output Codes

A Smoothed Boosting Algorithm Using Probabilistic Output Codes A Soothed Boosting Algorith Using Probabilistic Output Codes Rong Jin rongjin@cse.su.edu Dept. of Coputer Science and Engineering, Michigan State University, MI 48824, USA Jian Zhang jian.zhang@cs.cu.edu

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc

More information

How to Combine Tree-Search Methods in Reinforcement Learning

How to Combine Tree-Search Methods in Reinforcement Learning How to Cobine Tree-Searc Metods in Reinforceent Learning Yonatan Efroni Tecnion, Israel Gal Dalal Tecnion, Israel Bruno Scerrer INRIA, Villers les Nancy, France Sie Mannor Tecnion, Israel Abstract Finite-orizon

More information

arxiv: v3 [cs.ds] 22 Mar 2016

arxiv: v3 [cs.ds] 22 Mar 2016 A Shifting Bloo Filter Fraewor for Set Queries arxiv:1510.03019v3 [cs.ds] Mar 01 ABSTRACT Tong Yang Peing University, China yangtongeail@gail.co Yuanun Zhong Nanjing University, China un@sail.nju.edu.cn

More information

a a a a a a a m a b a b

a a a a a a a m a b a b Algebra / Trig Final Exa Study Guide (Fall Seester) Moncada/Dunphy Inforation About the Final Exa The final exa is cuulative, covering Appendix A (A.1-A.5) and Chapter 1. All probles will be ultiple choice

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

Introduction to Machine Learning. Recitation 11

Introduction to Machine Learning. Recitation 11 Introduction to Machine Learning Lecturer: Regev Schweiger Recitation Fall Seester Scribe: Regev Schweiger. Kernel Ridge Regression We now take on the task of kernel-izing ridge regression. Let x,...,

More information

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Proc. of the IEEE/OES Seventh Working Conference on Current Measureent Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Belinda Lipa Codar Ocean Sensors 15 La Sandra Way, Portola Valley, CA 98 blipa@pogo.co

More information

Ensemble Based on Data Envelopment Analysis

Ensemble Based on Data Envelopment Analysis Enseble Based on Data Envelopent Analysis So Young Sohn & Hong Choi Departent of Coputer Science & Industrial Systes Engineering, Yonsei University, Seoul, Korea Tel) 82-2-223-404, Fax) 82-2- 364-7807

More information

Bloom Filters. filters: A survey, Internet Mathematics, vol. 1 no. 4, pp , 2004.

Bloom Filters. filters: A survey, Internet Mathematics, vol. 1 no. 4, pp , 2004. Bloo Filters References A. Broder and M. Mitzenacher, Network applications of Bloo filters: A survey, Internet Matheatics, vol. 1 no. 4, pp. 485-509, 2004. Li Fan, Pei Cao, Jussara Aleida, Andrei Broder,

More information

MA455 Manifolds Solutions 1 May 2008

MA455 Manifolds Solutions 1 May 2008 MA455 Manifolds Solutions 1 May 2008 1. (i) Given real numbers a < b, find a diffeomorpism (a, b) R. Solution: For example first map (a, b) to (0, π/2) and ten map (0, π/2) diffeomorpically to R using

More information

LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION

LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LAURA EVANS.. Introduction Not all differential equations can be explicitly solved for y. Tis can be problematic if we need to know te value of y

More information

A NOTE ON BOOLEAN LATTICES AND FAREY SEQUENCES. Andrey O. Matveev Data-Center Co., RU , Ekaterinburg, P.O. Box 5, Russian Federation

A NOTE ON BOOLEAN LATTICES AND FAREY SEQUENCES. Andrey O. Matveev Data-Center Co., RU , Ekaterinburg, P.O. Box 5, Russian Federation INTEGERS: ELECTRONIC JOURNAL OF COMBINATORIAL NUMBER THEORY 7 27 #A2 A NOTE ON BOOLEAN LATTICES AND FAREY SEQUENCES Anrey O Matveev Data-Center Co RU-6234 Eaterinburg PO Box 5 Russian Feeration aoatveev@cru

More information

Testing Properties of Collections of Distributions

Testing Properties of Collections of Distributions Testing Properties of Collections of Distributions Reut Levi Dana Ron Ronitt Rubinfeld April 9, 0 Abstract We propose a fraework for studying property testing of collections of distributions, where the

More information

Estimating Parameters for a Gaussian pdf

Estimating Parameters for a Gaussian pdf Pattern Recognition and achine Learning Jaes L. Crowley ENSIAG 3 IS First Seester 00/0 Lesson 5 7 Noveber 00 Contents Estiating Paraeters for a Gaussian pdf Notation... The Pattern Recognition Proble...3

More information

How to Find the Derivative of a Function: Calculus 1

How to Find the Derivative of a Function: Calculus 1 Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te

More information

Using a De-Convolution Window for Operating Modal Analysis

Using a De-Convolution Window for Operating Modal Analysis Using a De-Convolution Window for Operating Modal Analysis Brian Schwarz Vibrant Technology, Inc. Scotts Valley, CA Mark Richardson Vibrant Technology, Inc. Scotts Valley, CA Abstract Operating Modal Analysis

More information

Ocean 420 Physical Processes in the Ocean Project 1: Hydrostatic Balance, Advection and Diffusion Answers

Ocean 420 Physical Processes in the Ocean Project 1: Hydrostatic Balance, Advection and Diffusion Answers Ocean 40 Physical Processes in the Ocean Project 1: Hydrostatic Balance, Advection and Diffusion Answers 1. Hydrostatic Balance a) Set all of the levels on one of the coluns to the lowest possible density.

More information

A = h w (1) Error Analysis Physics 141

A = h w (1) Error Analysis Physics 141 Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.

More information

HOMEWORK HELP 2 FOR MATH 151

HOMEWORK HELP 2 FOR MATH 151 HOMEWORK HELP 2 FOR MATH 151 Here we go; te second round of omework elp. If tere are oters you would like to see, let me know! 2.4, 43 and 44 At wat points are te functions f(x) and g(x) = xf(x)continuous,

More information

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy Storage Capacity and Dynaics of Nononotonic Networks Bruno Crespi a and Ignazio Lazzizzera b a. IRST, I-38050 Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I-38050 Povo (Trento) Italy INFN Gruppo

More information

The Algorithms Optimization of Artificial Neural Network Based on Particle Swarm

The Algorithms Optimization of Artificial Neural Network Based on Particle Swarm Send Orders for Reprints to reprints@benthascience.ae The Open Cybernetics & Systeics Journal, 04, 8, 59-54 59 Open Access The Algoriths Optiization of Artificial Neural Network Based on Particle Swar

More information