Supplementary Material: HCP: A Flexible CNN Framework for Multi-label Image Classification

Size: px
Start display at page:

Download "Supplementary Material: HCP: A Flexible CNN Framework for Multi-label Image Classification"

Transcription

1 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL.XX, NO.XX, Supplemetary Material: HCP: A Flexible CNN Framework for Multi-label Image Classificatio Yuchao Wei, Wei Xia, Mi Li, Jushi Huag, Bigbig Ni, Jia Dog, Yao Zhao, Seior Member, IEEE Shuicheg Ya, Seior Member, IEEE This documet provides detailed justificatios of model compoets for the results used i the paper. Basically, the experimetal results are based o VOC JUSTIFICATIONS OF IMAGE-FINE-TUNING AND HYPOTHESES-FINE-TUNING This justificatio is based o 10 selected hypotheses geerated by BING [2] for each traiig image. First, to demostrate the ecessity of image-fie-tuig, we compare with the method which does ot utilize image-fie-tuig to iitialize the last fully-coected layer. As show i Table 1, without image-fie-tuig, the performace will drop by 4%. The reaso may be explaied as follows. With the max poolig operatio, our method requires a reasoable predictio for each hypothesis at the begiig so that a advisable hypothesis ca be selected to optimize the parameters durig the back propagatio process. However, without image-fietuig, radom iitialized parameters for the last layer may geerate a irratioal predicted result for each hypothesis at the begiig. The, with the cross-patch max poolig, icorrect hypotheses may be selected to optimize the shared CNN, leadig to a local optimum for the optimizatio. TABLE 1 Compariso of HCP performace i terms of differet I-FT settigs o VOC 2007 with Alex Net [5] as the shared CNN ( i %). Without I-FT With I-FT Secod, to demostrate the ecessity of hypotheses-fie-tuig, we compare the cross-patch max-poolig result based o the image-fie-tuig model with that based o the hypotheses-fie-tuig model. From Table 2, we ca see that the improvemet from sigle image to 500 hypotheses based I-FT is limited, while the classificatio performace ca be sigificatly boosted with HCP. Therefore, hypotheses-fie-tuig ca further improve the classificatio accuracy. TABLE 2 Compariso i terms of 500 hypotheses for I-FT model ad HCP model o VOC 2007 with Alex Net as the shared CNN ( i %). I-FT HCP Sigle Image hypotheses JUSTIFICATION OF LOSS FUNCTION To validate the effectiveess of the squared loss fuctio utilized i the HCP framework, we compare it with other multi-label loss fuctios. Four loss fuctios, i.e., cross-etropy loss [3], hige loss [1], label-wise crossetropy loss ad pairwise rakig loss [3], are maily employed to make the compariso. Details of loss fuctios are show i the followig.

2 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL.XX, NO.XX, Details of Loss Fuctios Deote X = {x i i = 1,..., } as the image traiig set. Similar to the otatios defied i [3], we deote the deep CNN by f( ) where all the layers filter the give image x i. The f( ) produces a c-dimesioal output of activatios, where c is the umber of the predefied labels Cross-etropy Loss The softmax fuctio is used to compute the posterior probability of a give image x i belogig to the class j, i.e. exp(f j (x i )) p ij = P (j x i ) = c k=1 exp(f (1) k(x i )) where f j (x i ) is the j th activatio value for the image x i. Let y i = [y i1, y i2,, y ic ] be the label vector of the i th image, where y ij = 1 (j = 1,, c) if the image is aotated by the class j, ad y ij = 0 otherwise. The groud-truth probability vector of the i th image is defied as ˆp i = y i / y i 1. Give the etwork predictio i Eq. ( 1), the loss fuctio is the defied as J = 1 c ˆp ij log(p ij ) = 1 i=1 j=1 1 log(p ij ) c i=1 i,+ j C + where c i,+ is the umber of the groud-truth labels for the i th image, ad C + deotes the idex set of the labeled classes Pairwise Rakig Loss This loss fuctio is usually utilized for the aotatio problem. Specifically, the target of this loss fuctio is to rak the positive labels to have higher scores tha egative labels. The loss fuctio is the defied as J = 1 max(0, 1 f j (x i ) + f k (x i )) (3) i=1 k C j C + where C + ad C deote the idex sets of labeled ad o-labeled classes, respectively Hige Loss The hige loss is used for maximum-margi classificatio, which is most otably for support vector machies (SVM). Idicated by [1], the loss fuctio is defied as J = 1 c max (0, 1 y ij f j (x i )) 2 (4) where i=1 j=1 { +1 if xi is aotated with the j y ij = th label 1 otherwise Label-wise Cross-etropy Loss We preset a ew loss fuctio for multi-label classificatio, which is called label-wise cross-etropy (LWCE) loss. The LWCE loss splits the traditioal cross-etropy loss ito multiple oe vs. all cross-etropy losses for each class. Specifically, we use the deep CNN as the predictio mappig ad chage the output dimesio of the last fully-coected layer of the etwork from c to 2c. The two succeedig dimesios are the fed to the softmax fuctio to compute the cross-etropy loss for each class. Deote ˆf j ( ) R 2 as the 2 dimesioal output of activatios for the j th (j = 1 c) label. We formulate the posterior probability of the give image x i belogig to the j th label as exp( p ij = P (j x i ) = ˆf j1 (x i )) 2 k=1 exp( ˆf (5) jk (x i )) where ˆf jk is the j th dimesio of ˆf j for k {1, 2}. (2)

3 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL.XX, NO.XX, Suppose y ij = [y ij1, y ij2 ] is the 2-dimesioal label vector of the j th class, where y ij1 = 1, y ij2 = 0 if the image is aotated with class j, ad otherwise y ij1 = 0, y ij2 = 1. The, the loss fuctio to be miimized ca be writte as J = 1 c 2 y ijk log(p ijk ) i=1 j=1 k=1 = 1 (6) log(p ij1 )+ log(p ij2 ) i=1 j C + j C where C + ad C deote the idex sets of labeled ad o-labeled classes, respectively. 2.2 Results First, from the experimetal results as show i Table 3, it ca be observed that the squared loss ad the cross-etropy loss reach better results tha other loss fuctios. The uderlyig reaso may be explaied as follows. With the softmax operatio across all labels, the cotext relatioship amog labels has bee implicitly utilized (cocurret or exclusive) i squared loss ad cross-etropy loss, which is valuable for multi-label image classificatio as for VOC datasets. However, for both hige loss ad label-wise cross-etropy loss, the optimizatio for each class is idepedet, where the cotext iformatio of other labels i ot take ito cosideratio. Secod, experimetal result based o pairwise rakig loss shows that it is less effective tha other loss fuctios. The target of this loss fuctio is to rak the positive labels with higher predicted scores tha egative labels withi a particular sample. It however caot guaratee that positive samples are raked higher tha egative samples for a specific class. Therefore, its performace may be worse. Third, the experimets show that squared loss is slightly better tha cross-etropy loss, ad the uderlyig reaso may be as follows. As ca be see from the mathematical defiitio i Eq. (2), the cross-etropy loss oly aims to promote the predicted score(s) o groud-truth label(s) ad does ot care about the predicted scores o the egative labels, ad thus it is possible that a particular egative label may also have relatively a large value, which may degrade the multi-label classificatio performace (cofidece value is already ot too big whe the umber of positive labels is larger tha oe). However, squared loss attempts to suppress the predicted scores of all the egative labels to zero. Also miimizig the squared loss mathematically shall ecourage the squared losses for all idividual egative labels of a image to be similar to each other, amely balaced, which may beefit the classificatio accuracy sice there will be o big cofideces for all egative labels. All these experimets are based o 10 selected hypotheses (geerated by BING) for traiig ad 500 hypotheses for testig. TABLE 3 Compariso of HCP performace i terms of differet loss fuctios o VOC 2007 with Alex Net as the shared CNN ( i %). Loss Fuctio Squared Loss 81.5 Cross-etropy Loss [3] 80.2 Hige Loss [1] 79.9 Label-wise Cross-etropy Loss 78.9 Pairwise Rakig Loss [3] JUSTIFICATION OF HYPOTHESIS GENERATORS To validate the effectiveess of usig object proposal method, two other schemes, i.e., radom crops ad uiform crops, are itroduced for compariso. For the radom crops, we estimate a multivariate Gaussia distributio for the boudig box ceter positio, square root area, ad log aspect ratio. After calculatig mea ad covariace o the traiig set, we sample proposals by followig this distributio. For the uiform corps, we uiformly sample the boudig box ceter positio, square root area, ad log aspect ratio. These two cropped methods are cosistet with those proposed i [4]. We utilize the codes released by [4] to geerate the correspodig boudig boxes for each image. Sice o objectess cofidece score is assiged to each boudig box, we radomly select oe hypothesis from each cluster ceter for traiig. I additio, to evaluate

4 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL.XX, NO.XX, how other better yet more complex proposal geeratio methods affect the performace compared with BING, we add ew experimets based o aother proposal geeratio method, EdgeBoxes [7], due to its acceptable speed ad high quality of proposals. All these experimets are based o 10 selected hypotheses for the traiig of HCP. Top-500 hypotheses geerated by the correspodig geerators (i.e., Uiform, Radom, BING ad Edgeboxes) are used for testig. Table 4 shows the compariso results o the VOC 2007 test set. It ca be observed that our HCP framework still works well based o uiform/radom crops. However, the results of these two schemes are worse tha those proposal-geeratio schemes, i.e., BING ad EdgeBoxes. The uderlyig reaso is that proposal geerator ca cover or recall hypotheses with more complete ad diverse objects compared with radom/uiform crops. TABLE 4 Compariso of HCP performace i terms of differet hypotheses selectio strategies o VOC 2007 with Alex Net as the shared CNN ( i %). Hypotheses Selectio Uiform 80.6 Radom 80.9 BING 81.5 Edgeboxes 82.3 TABLE 5 Compariso of HCP performace i terms of differet umber of clusters ad proposal methods o VOC 2007 ( i %). Alex Net VGG Net BING EdgeBoxes BING EdgeBoxes 5-clusters clusters clusters JUSTIFICATION OF THE NUMBER OF CLUSTERS We vary the umber of clusters durig the HCP traiig stage to observe its ifluece based o two proposal methods, i.e., BING [2] ad EdgeBoxes [7]. The experimets are addressed based o -clusters (=5,10,15), where is the umber of ceters durig the clusterig operatio. The compariso results are show i Table 5 ad all the testig results are based o top-500 hypotheses for each image. From Table 5, it ca be observed that the performace of HCP ca be boosted with the icrease of the cluster umber. This is because more clusters ca brig higher coverage rate for objects of traiig images. I additio, traiig as well as testig with hypotheses produced by Edgeboxes performs better tha that of BING. The mai reaso is the quality of the hypotheses. As idicated i [4], hypotheses geerated by Edgeboxes are better tha those produced by BING. It takes about 1 secod to cluster the top-500 boudig boxes geerated by BING or EdgeBoxes for traiig. Based o the top-500 hypotheses, the testig time of Alex Net ad VGG Net [6] is about 3s/image ad 10s/image o a GPU icludig the time of proposal geeratio (BING: 0.2s/image, EdgeBoxes: 0.25s/image). Details of all experimetal justificatios o VOC 2007 are show i Table 6. REFERENCES [1] K. Chatfield, K. Simoya, A. Vedaldi, ad A. Zisserma. Retur of the devil i the details: Delvig deep ito covolutioal ets. I British Machie Visio Coferece, [2] M.-M. Cheg, Z. Zhag, W.-Y. Li, ad P. H. S. Torr. BING: Biarized ormed gradiets for objectess estimatio at 300fps. I Computer Visio ad Patter Recogitio, [3] Y. Gog, Y. Jia, T. K. leug, A. Toshev, ad S. Ioffe. deep covolutioal rakig for multi label image aotatio. I Iteratioal Coferece o Learig Represetatios, [4] J. Hosag, R. Beeso, P. Dollár, ad B. Schiele. What makes for effective detectio proposals? arxiv preprit arxiv: , [5] A. Krizhevsky, I. Sutskever, ad G. Hito. Imageet classificatio with deep covolutioal eural etworks. I Neural Iformatio Processig Systems, pages , [6] K. Simoya ad A. Zisserma. Very deep covolutioal etworks for large-scale image recogitio. arxiv preprit arxiv: , [7] C. L. Zitick ad P. Dollár. Edge boxes: Locatig object proposals from edges. I Europea Coferece o Computer Visio, pages Spriger, 2014.

5 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL.XX, NO.XX, TABLE 6 Details of experimetal results o VOC plae bike bird boat bottle bus car cat chair Justificatio of loss fuctios (Traiig: 10 clusters, Testig: Top-500 hypotheses, Proposal method: BING, Shared CNN: Alex Net) : Squared Cross-etropy Hige Label-wise Pairwise Rakig Uiform/Radom croppig vs. proposal methods (Traiig: 10 clusters, Testig: Top-500 hypotheses, Shared CNN: Alex Net) : Uiform Radom BING EdgeBoxes Justificatio of proposal methods based o 5 clusters (Testig: Top-500 hypotheses): Alex+BING Alex+EdgeBoxes VGG+BING VGG+EdgeBoxes Justificatio of proposal methods based o 10 clusters (Testig: Top-500 hypotheses): Alex+BING Alex+EdgeBoxes VGG+BING VGG+EdgeBoxes Justificatio of proposal methods based o 15 clusters (Testig: Top-500 hypotheses): Alex+BING Alex+EdgeBoxes VGG+BING VGG+EdgeBoxes cow table dog horse motor perso plat sheep sofa trai tv

10-701/ Machine Learning Mid-term Exam Solution

10-701/ Machine Learning Mid-term Exam Solution 0-70/5-78 Machie Learig Mid-term Exam Solutio Your Name: Your Adrew ID: True or False (Give oe setece explaatio) (20%). (F) For a cotiuous radom variable x ad its probability distributio fuctio p(x), it

More information

1 Review of Probability & Statistics

1 Review of Probability & Statistics 1 Review of Probability & Statistics a. I a group of 000 people, it has bee reported that there are: 61 smokers 670 over 5 960 people who imbibe (drik alcohol) 86 smokers who imbibe 90 imbibers over 5

More information

Introduction to Machine Learning DIS10

Introduction to Machine Learning DIS10 CS 189 Fall 017 Itroductio to Machie Learig DIS10 1 Fu with Lagrage Multipliers (a) Miimize the fuctio such that f (x,y) = x + y x + y = 3. Solutio: The Lagragia is: L(x,y,λ) = x + y + λ(x + y 3) Takig

More information

Expectation-Maximization Algorithm.

Expectation-Maximization Algorithm. Expectatio-Maximizatio Algorithm. Petr Pošík Czech Techical Uiversity i Prague Faculty of Electrical Egieerig Dept. of Cyberetics MLE 2 Likelihood.........................................................................................................

More information

Machine Learning Theory (CS 6783)

Machine Learning Theory (CS 6783) Machie Learig Theory (CS 6783) Lecture 2 : Learig Frameworks, Examples Settig up learig problems. X : istace space or iput space Examples: Computer Visio: Raw M N image vectorized X = 0, 255 M N, SIFT

More information

Information-based Feature Selection

Information-based Feature Selection Iformatio-based Feature Selectio Farza Faria, Abbas Kazeroui, Afshi Babveyh Email: {faria,abbask,afshib}@staford.edu 1 Itroductio Feature selectio is a topic of great iterest i applicatios dealig with

More information

Mixtures of Gaussians and the EM Algorithm

Mixtures of Gaussians and the EM Algorithm Mixtures of Gaussias ad the EM Algorithm CSE 6363 Machie Learig Vassilis Athitsos Computer Sciece ad Egieerig Departmet Uiversity of Texas at Arligto 1 Gaussias A popular way to estimate probability desity

More information

Empirical Process Theory and Oracle Inequalities

Empirical Process Theory and Oracle Inequalities Stat 928: Statistical Learig Theory Lecture: 10 Empirical Process Theory ad Oracle Iequalities Istructor: Sham Kakade 1 Risk vs Risk See Lecture 0 for a discussio o termiology. 2 The Uio Boud / Boferoi

More information

Selective Prediction

Selective Prediction COMS 6998-4 Fall 2017 November 8, 2017 Selective Predictio Preseter: Rog Zhou Scribe: Wexi Che 1 Itroductio I our previous discussio o a variatio o the Valiat Model [3], the described learer has the ability

More information

Optimally Sparse SVMs

Optimally Sparse SVMs A. Proof of Lemma 3. We here prove a lower boud o the umber of support vectors to achieve geeralizatio bouds of the form which we cosider. Importatly, this result holds ot oly for liear classifiers, but

More information

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 12

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 12 Machie Learig Theory Tübige Uiversity, WS 06/07 Lecture Tolstikhi Ilya Abstract I this lecture we derive risk bouds for kerel methods. We will start by showig that Soft Margi kerel SVM correspods to miimizig

More information

Estimation for Complete Data

Estimation for Complete Data Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of

More information

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015 ECE 8527: Itroductio to Machie Learig ad Patter Recogitio Midterm # 1 Vaishali Ami Fall, 2015 tue39624@temple.edu Problem No. 1: Cosider a two-class discrete distributio problem: ω 1 :{[0,0], [2,0], [2,2],

More information

6.867 Machine learning, lecture 7 (Jaakkola) 1

6.867 Machine learning, lecture 7 (Jaakkola) 1 6.867 Machie learig, lecture 7 (Jaakkola) 1 Lecture topics: Kerel form of liear regressio Kerels, examples, costructio, properties Liear regressio ad kerels Cosider a slightly simpler model where we omit

More information

Topics Machine learning: lecture 2. Review: the learning problem. Hypotheses and estimation. Estimation criterion cont d. Estimation criterion

Topics Machine learning: lecture 2. Review: the learning problem. Hypotheses and estimation. Estimation criterion cont d. Estimation criterion .87 Machie learig: lecture Tommi S. Jaakkola MIT CSAIL tommi@csail.mit.edu Topics The learig problem hypothesis class, estimatio algorithm loss ad estimatio criterio samplig, empirical ad epected losses

More information

6.867 Machine learning

6.867 Machine learning 6.867 Machie learig Mid-term exam October, ( poits) Your ame ad MIT ID: Problem We are iterested here i a particular -dimesioal liear regressio problem. The dataset correspodig to this problem has examples

More information

Clustering. CM226: Machine Learning for Bioinformatics. Fall Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar.

Clustering. CM226: Machine Learning for Bioinformatics. Fall Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar. Clusterig CM226: Machie Learig for Bioiformatics. Fall 216 Sriram Sakararama Ackowledgmets: Fei Sha, Ameet Talwalkar Clusterig 1 / 42 Admiistratio HW 1 due o Moday. Email/post o CCLE if you have questios.

More information

Chapter 7. Support Vector Machine

Chapter 7. Support Vector Machine Chapter 7 Support Vector Machie able of Cotet Margi ad support vectors SVM formulatio Slack variables ad hige loss SVM for multiple class SVM ith Kerels Relevace Vector Machie Support Vector Machie (SVM)

More information

Linear Support Vector Machines

Linear Support Vector Machines Liear Support Vector Machies David S. Roseberg The Support Vector Machie For a liear support vector machie (SVM), we use the hypothesis space of affie fuctios F = { f(x) = w T x + b w R d, b R } ad evaluate

More information

Hypothesis Testing. Evaluation of Performance of Learned h. Issues. Trade-off Between Bias and Variance

Hypothesis Testing. Evaluation of Performance of Learned h. Issues. Trade-off Between Bias and Variance Hypothesis Testig Empirically evaluatig accuracy of hypotheses: importat activity i ML. Three questios: Give observed accuracy over a sample set, how well does this estimate apply over additioal samples?

More information

Vassilis Katsouros, Vassilis Papavassiliou and Christos Emmanouilidis

Vassilis Katsouros, Vassilis Papavassiliou and Christos Emmanouilidis Vassilis Katsouros, Vassilis Papavassiliou ad Christos Emmaouilidis ATHENA Research & Iovatio Cetre, Greece www.athea-iovatio.gr www.ceti.athea-iovatio.gr/compsys e-mail: christosem AT ieee.org Problem

More information

Frequentist Inference

Frequentist Inference Frequetist Iferece The topics of the ext three sectios are useful applicatios of the Cetral Limit Theorem. Without kowig aythig about the uderlyig distributio of a sequece of radom variables {X i }, for

More information

Pixel Recurrent Neural Networks

Pixel Recurrent Neural Networks Pixel Recurret Neural Networks Aa ro va de Oord, Nal Kalchbreer, Koray Kavukcuoglu Google DeepMid August 2016 Preseter - Neha M Example problem (completig a image) Give the first half of the image, create

More information

Properties and Hypothesis Testing

Properties and Hypothesis Testing Chapter 3 Properties ad Hypothesis Testig 3.1 Types of data The regressio techiques developed i previous chapters ca be applied to three differet kids of data. 1. Cross-sectioal data. 2. Time series data.

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

Machine Learning. Ilya Narsky, Caltech

Machine Learning. Ilya Narsky, Caltech Machie Learig Ilya Narsky, Caltech Lecture 4 Multi-class problems. Multi-class versios of Neural Networks, Decisio Trees, Support Vector Machies ad AdaBoost. Reductio of a multi-class problem to a set

More information

Pattern Classification, Ch4 (Part 1)

Pattern Classification, Ch4 (Part 1) Patter Classificatio All materials i these slides were take from Patter Classificatio (2d ed) by R O Duda, P E Hart ad D G Stork, Joh Wiley & Sos, 2000 with the permissio of the authors ad the publisher

More information

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 3

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 3 Machie Learig Theory Tübige Uiversity, WS 06/07 Lecture 3 Tolstikhi Ilya Abstract I this lecture we will prove the VC-boud, which provides a high-probability excess risk boud for the ERM algorithm whe

More information

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10 DS 00: Priciples ad Techiques of Data Sciece Date: April 3, 208 Name: Hypothesis Testig Discussio #0. Defie these terms below as they relate to hypothesis testig. a) Data Geeratio Model: Solutio: A set

More information

Lecture 7: October 18, 2017

Lecture 7: October 18, 2017 Iformatio ad Codig Theory Autum 207 Lecturer: Madhur Tulsiai Lecture 7: October 8, 207 Biary hypothesis testig I this lecture, we apply the tools developed i the past few lectures to uderstad the problem

More information

1 Review and Overview

1 Review and Overview DRAFT a fial versio will be posted shortly CS229T/STATS231: Statistical Learig Theory Lecturer: Tegyu Ma Lecture #3 Scribe: Migda Qiao October 1, 2013 1 Review ad Overview I the first half of this course,

More information

Week 1, Lecture 2. Neural Network Basics. Announcements: HW 1 Due on 10/8 Data sets for HW 1 are online Project selection 10/11. Suggested reading :

Week 1, Lecture 2. Neural Network Basics. Announcements: HW 1 Due on 10/8 Data sets for HW 1 are online Project selection 10/11. Suggested reading : ME 537: Learig-Based Cotrol Week 1, Lecture 2 Neural Network Basics Aoucemets: HW 1 Due o 10/8 Data sets for HW 1 are olie Proect selectio 10/11 Suggested readig : NN survey paper (Zhag Chap 1, 2 ad Sectios

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Patter Recogitio Classificatio: No-Parametric Modelig Hamid R. Rabiee Jafar Muhammadi Sprig 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Ageda Parametric Modelig No-Parametric Modelig

More information

Binary classification, Part 1

Binary classification, Part 1 Biary classificatio, Part 1 Maxim Ragisky September 25, 2014 The problem of biary classificatio ca be stated as follows. We have a radom couple Z = (X,Y ), where X R d is called the feature vector ad Y

More information

Vector Quantization: a Limiting Case of EM

Vector Quantization: a Limiting Case of EM . Itroductio & defiitios Assume that you are give a data set X = { x j }, j { 2,,, }, of d -dimesioal vectors. The vector quatizatio (VQ) problem requires that we fid a set of prototype vectors Z = { z

More information

Machine Learning Brett Bernstein

Machine Learning Brett Bernstein Machie Learig Brett Berstei Week 2 Lecture: Cocept Check Exercises Starred problems are optioal. Excess Risk Decompositio 1. Let X = Y = {1, 2,..., 10}, A = {1,..., 10, 11} ad suppose the data distributio

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

Improved Listwise Collaborative Filtering with High- Rating-Based Similarity and Temporary Ratings

Improved Listwise Collaborative Filtering with High- Rating-Based Similarity and Temporary Ratings Improved Listwise Collaborative Filterig with High- Ratig-Based Similarity ad Temporary Ratigs Yoshiki Tsuchiya Uiversity of Tsukuba Tsukuba, Japa tsuchiya@cmu.iit.tsukuba.ac.jp ABSTRACT I this paper,

More information

Solution of Final Exam : / Machine Learning

Solution of Final Exam : / Machine Learning Solutio of Fial Exam : 10-701/15-781 Machie Learig Fall 2004 Dec. 12th 2004 Your Adrew ID i capital letters: Your full ame: There are 9 questios. Some of them are easy ad some are more difficult. So, if

More information

ENGI 4421 Confidence Intervals (Two Samples) Page 12-01

ENGI 4421 Confidence Intervals (Two Samples) Page 12-01 ENGI 44 Cofidece Itervals (Two Samples) Page -0 Two Sample Cofidece Iterval for a Differece i Populatio Meas [Navidi sectios 5.4-5.7; Devore chapter 9] From the cetral limit theorem, we kow that, for sufficietly

More information

Problem Set 4 Due Oct, 12

Problem Set 4 Due Oct, 12 EE226: Radom Processes i Systems Lecturer: Jea C. Walrad Problem Set 4 Due Oct, 12 Fall 06 GSI: Assae Gueye This problem set essetially reviews detectio theory ad hypothesis testig ad some basic otios

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

Machine Learning for Data Science (CS 4786)

Machine Learning for Data Science (CS 4786) Machie Learig for Data Sciece CS 4786) Lecture & 3: Pricipal Compoet Aalysis The text i black outlies high level ideas. The text i blue provides simple mathematical details to derive or get to the algorithm

More information

(all terms are scalars).the minimization is clearer in sum notation:

(all terms are scalars).the minimization is clearer in sum notation: 7 Multiple liear regressio: with predictors) Depedet data set: y i i = 1, oe predictad, predictors x i,k i = 1,, k = 1, ' The forecast equatio is ŷ i = b + Use matrix otatio: k =1 b k x ik Y = y 1 y 1

More information

The Expectation-Maximization (EM) Algorithm

The Expectation-Maximization (EM) Algorithm The Expectatio-Maximizatio (EM) Algorithm Readig Assigmets T. Mitchell, Machie Learig, McGraw-Hill, 997 (sectio 6.2, hard copy). S. Gog et al. Dyamic Visio: From Images to Face Recogitio, Imperial College

More information

Grouping 2: Spectral and Agglomerative Clustering. CS 510 Lecture #16 April 2 nd, 2014

Grouping 2: Spectral and Agglomerative Clustering. CS 510 Lecture #16 April 2 nd, 2014 Groupig 2: Spectral ad Agglomerative Clusterig CS 510 Lecture #16 April 2 d, 2014 Groupig (review) Goal: Detect local image features (SIFT) Describe image patches aroud features SIFT, SURF, HoG, LBP, Group

More information

10/2/ , 5.9, Jacob Hays Amit Pillay James DeFelice

10/2/ , 5.9, Jacob Hays Amit Pillay James DeFelice 0//008 Liear Discrimiat Fuctios Jacob Hays Amit Pillay James DeFelice 5.8, 5.9, 5. Miimum Squared Error Previous methods oly worked o liear separable cases, by lookig at misclassified samples to correct

More information

Discrete-Time Systems, LTI Systems, and Discrete-Time Convolution

Discrete-Time Systems, LTI Systems, and Discrete-Time Convolution EEL5: Discrete-Time Sigals ad Systems. Itroductio I this set of otes, we begi our mathematical treatmet of discrete-time s. As show i Figure, a discrete-time operates or trasforms some iput sequece x [

More information

This is an introductory course in Analysis of Variance and Design of Experiments.

This is an introductory course in Analysis of Variance and Design of Experiments. 1 Notes for M 384E, Wedesday, Jauary 21, 2009 (Please ote: I will ot pass out hard-copy class otes i future classes. If there are writte class otes, they will be posted o the web by the ight before class

More information

Electricity consumption forecasting method based on MPSO-BP neural network model Youshan Zhang 1, 2,a, Liangdong Guo2, b,qi Li 3, c and Junhui Li2, d

Electricity consumption forecasting method based on MPSO-BP neural network model Youshan Zhang 1, 2,a, Liangdong Guo2, b,qi Li 3, c and Junhui Li2, d 4th Iteratioal Coferece o Electrical & Electroics Egieerig ad Computer Sciece (ICEEECS 2016) Electricity cosumptio forecastig method based o eural etwork model Yousha Zhag 1, 2,a, Liagdog Guo2, b,qi Li

More information

Inferential Statistics. Inference Process. Inferential Statistics and Probability a Holistic Approach. Inference Process.

Inferential Statistics. Inference Process. Inferential Statistics and Probability a Holistic Approach. Inference Process. Iferetial Statistics ad Probability a Holistic Approach Iferece Process Chapter 8 Poit Estimatio ad Cofidece Itervals This Course Material by Maurice Geraghty is licesed uder a Creative Commos Attributio-ShareAlike

More information

Measures of Spread: Standard Deviation

Measures of Spread: Standard Deviation Measures of Spread: Stadard Deviatio So far i our study of umerical measures used to describe data sets, we have focused o the mea ad the media. These measures of ceter tell us the most typical value of

More information

A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS

A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS J. Japa Statist. Soc. Vol. 41 No. 1 2011 67 73 A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS Yoichi Nishiyama* We cosider k-sample ad chage poit problems for idepedet data i a

More information

Outline. Linear regression. Regularization functions. Polynomial curve fitting. Stochastic gradient descent for regression. MLE for regression

Outline. Linear regression. Regularization functions. Polynomial curve fitting. Stochastic gradient descent for regression. MLE for regression REGRESSION 1 Outlie Liear regressio Regularizatio fuctios Polyomial curve fittig Stochastic gradiet descet for regressio MLE for regressio Step-wise forward regressio Regressio methods Statistical techiques

More information

New Particle Swarm Neural Networks Model Based Long Term Electrical Load Forecasting in Slovakia

New Particle Swarm Neural Networks Model Based Long Term Electrical Load Forecasting in Slovakia New Particle Swarm Neural Networks Model Based Log Term Electrical Load Forecastig i Slovakia S.H. OUDJANA 1, A. HELLAL 2 1, 2 Uiversity of Laghouat, Departmet of Electrical Egieerig, Laboratory of Aalysis

More information

Outline. CSCI-567: Machine Learning (Spring 2019) Outline. Prof. Victor Adamchik. Mar. 26, 2019

Outline. CSCI-567: Machine Learning (Spring 2019) Outline. Prof. Victor Adamchik. Mar. 26, 2019 Outlie CSCI-567: Machie Learig Sprig 209 Gaussia mixture models Prof. Victor Adamchik 2 Desity estimatio U of Souther Califoria Mar. 26, 209 3 Naive Bayes Revisited March 26, 209 / 57 March 26, 209 2 /

More information

Simulation. Two Rule For Inverting A Distribution Function

Simulation. Two Rule For Inverting A Distribution Function Simulatio Two Rule For Ivertig A Distributio Fuctio Rule 1. If F(x) = u is costat o a iterval [x 1, x 2 ), the the uiform value u is mapped oto x 2 through the iversio process. Rule 2. If there is a jump

More information

A New Multivariate Markov Chain Model with Applications to Sales Demand Forecasting

A New Multivariate Markov Chain Model with Applications to Sales Demand Forecasting Iteratioal Coferece o Idustrial Egieerig ad Systems Maagemet IESM 2007 May 30 - Jue 2 BEIJING - CHINA A New Multivariate Markov Chai Model with Applicatios to Sales Demad Forecastig Wai-Ki CHING a, Li-Mi

More information

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen)

Goodness-of-Fit Tests and Categorical Data Analysis (Devore Chapter Fourteen) Goodess-of-Fit Tests ad Categorical Data Aalysis (Devore Chapter Fourtee) MATH-252-01: Probability ad Statistics II Sprig 2019 Cotets 1 Chi-Squared Tests with Kow Probabilities 1 1.1 Chi-Squared Testig................

More information

Lecture 10: Performance Evaluation of ML Methods

Lecture 10: Performance Evaluation of ML Methods CSE57A Machie Learig Sprig 208 Lecture 0: Performace Evaluatio of ML Methods Istructor: Mario Neuma Readig: fcml: 5.4 (Performace); esl: 7.0 (Cross-Validatio); optioal book: Evaluatio Learig Algorithms

More information

Support vector machine revisited

Support vector machine revisited 6.867 Machie learig, lecture 8 (Jaakkola) 1 Lecture topics: Support vector machie ad kerels Kerel optimizatio, selectio Support vector machie revisited Our task here is to first tur the support vector

More information

Efficient GMM LECTURE 12 GMM II

Efficient GMM LECTURE 12 GMM II DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet

More information

Naïve Bayes. Naïve Bayes

Naïve Bayes. Naïve Bayes Statistical Data Miig ad Machie Learig Hilary Term 206 Dio Sejdiovic Departmet of Statistics Oxford Slides ad other materials available at: http://www.stats.ox.ac.uk/~sejdiov/sdmml : aother plug-i classifier

More information

Technical Report 670 STATISTICS DEPARTMENT UNIVERSITY OF CALIFORNIA AT BERKELEY. September 9, 2004

Technical Report 670 STATISTICS DEPARTMENT UNIVERSITY OF CALIFORNIA AT BERKELEY. September 9, 2004 CONSISTENCY FOR A SIMPLE MODEL OF RANDOM FORESTS Leo Breima Techical Report 670 STATISTICS DEPARTMENT UNIVERSITY OF CALIFORNIA AT BERKELEY ^: 0) Itroductio September 9, 2004 Radom Forests is a classificatio

More information

Optimization Methods MIT 2.098/6.255/ Final exam

Optimization Methods MIT 2.098/6.255/ Final exam Optimizatio Methods MIT 2.098/6.255/15.093 Fial exam Date Give: December 19th, 2006 P1. [30 pts] Classify the followig statemets as true or false. All aswers must be well-justified, either through a short

More information

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 5 STATISTICS II. Mea ad stadard error of sample data. Biomial distributio. Normal distributio 4. Samplig 5. Cofidece itervals

More information

Linear Classifiers III

Linear Classifiers III Uiversität Potsdam Istitut für Iformatik Lehrstuhl Maschielles Lere Liear Classifiers III Blaie Nelso, Tobias Scheffer Cotets Classificatio Problem Bayesia Classifier Decisio Liear Classifiers, MAP Models

More information

1 of 7 7/16/2009 6:06 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 6. Order Statistics Defiitios Suppose agai that we have a basic radom experimet, ad that X is a real-valued radom variable

More information

ME 539, Fall 2008: Learning-Based Control

ME 539, Fall 2008: Learning-Based Control ME 539, Fall 2008: Learig-Based Cotrol Neural Network Basics 10/1/2008 & 10/6/2008 Uiversity Orego State Neural Network Basics Questios??? Aoucemet: Homework 1 has bee posted Due Friday 10/10/08 at oo

More information

CHAPTER 4 BIVARIATE DISTRIBUTION EXTENSION

CHAPTER 4 BIVARIATE DISTRIBUTION EXTENSION CHAPTER 4 BIVARIATE DISTRIBUTION EXTENSION 4. Itroductio Numerous bivariate discrete distributios have bee defied ad studied (see Mardia, 97 ad Kocherlakota ad Kocherlakota, 99) based o various methods

More information

Intro to Learning Theory

Intro to Learning Theory Lecture 1, October 18, 2016 Itro to Learig Theory Ruth Urer 1 Machie Learig ad Learig Theory Comig soo 2 Formal Framework 21 Basic otios I our formal model for machie learig, the istaces to be classified

More information

Output Analysis (2, Chapters 10 &11 Law)

Output Analysis (2, Chapters 10 &11 Law) B. Maddah ENMG 6 Simulatio Output Aalysis (, Chapters 10 &11 Law) Comparig alterative system cofiguratio Sice the output of a simulatio is radom, the comparig differet systems via simulatio should be doe

More information

Deep Neural Networks CMSC 422 MARINE CARPUAT. Deep learning slides credit: Vlad Morariu

Deep Neural Networks CMSC 422 MARINE CARPUAT. Deep learning slides credit: Vlad Morariu Deep Neural Networks CMSC 422 MARINE CARPUAT marie@cs.umd.edu Deep learig slides credit: Vlad Morariu Traiig (Deep) Neural Networks Computatioal graphs Improvemets to gradiet descet Stochastic gradiet

More information

Provläsningsexemplar / Preview TECHNICAL REPORT INTERNATIONAL SPECIAL COMMITTEE ON RADIO INTERFERENCE

Provläsningsexemplar / Preview TECHNICAL REPORT INTERNATIONAL SPECIAL COMMITTEE ON RADIO INTERFERENCE TECHNICAL REPORT CISPR 16-4-3 2004 AMENDMENT 1 2006-10 INTERNATIONAL SPECIAL COMMITTEE ON RADIO INTERFERENCE Amedmet 1 Specificatio for radio disturbace ad immuity measurig apparatus ad methods Part 4-3:

More information

CHAPTER I: Vector Spaces

CHAPTER I: Vector Spaces CHAPTER I: Vector Spaces Sectio 1: Itroductio ad Examples This first chapter is largely a review of topics you probably saw i your liear algebra course. So why cover it? (1) Not everyoe remembers everythig

More information

Linear Associator Linear Layer

Linear Associator Linear Layer Hebbia Learig opic 6 Note: lecture otes by Michael Negevitsky (uiversity of asmaia) Bob Keller (Harvey Mudd College CA) ad Marti Haga (Uiversity of Colorado) are used Mai idea: learig based o associatio

More information

Classification of fragile states based on machine learning

Classification of fragile states based on machine learning Classificatio of fragile states based o machie learig Yuqua Li 1, Hehua Yao 1 1 Departmet of Computer Techology ad Applicatio, Qighai Uiversity, Xiig, Chia Abstract. The study of fragile states has become

More information

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara

Econ 325 Notes on Point Estimator and Confidence Interval 1 By Hiro Kasahara Poit Estimator Eco 325 Notes o Poit Estimator ad Cofidece Iterval 1 By Hiro Kasahara Parameter, Estimator, ad Estimate The ormal probability desity fuctio is fully characterized by two costats: populatio

More information

Multilayer perceptrons

Multilayer perceptrons Multilayer perceptros If traiig set is ot liearly separable, a etwork of McCulloch-Pitts uits ca give a solutio If o loop exists i etwork, called a feedforward etwork (else, recurret etwork) A two-layer

More information

There is no straightforward approach for choosing the warmup period l.

There is no straightforward approach for choosing the warmup period l. B. Maddah INDE 504 Discrete-Evet Simulatio Output Aalysis () Statistical Aalysis for Steady-State Parameters I a otermiatig simulatio, the iterest is i estimatig the log ru steady state measures of performace.

More information

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification INF 4300 90 Itroductio to classifictio Ae Solberg ae@ifiuioo Based o Chapter -6 i Duda ad Hart: atter Classificatio 90 INF 4300 Madator proect Mai task: classificatio You must implemet a classificatio

More information

An Information-Theoretic View for Deep Learning

An Information-Theoretic View for Deep Learning A Iformatio-Theoretic View for Deep Learig Jigwei Zhag Togliag Liu Dacheg Tao arxiv:804.09060v8 stat.ml Oct 08 Abstract Deep learig has trasformed computer visio, atural laguage processig, ad speech recogitio,

More information

Final Examination Solutions 17/6/2010

Final Examination Solutions 17/6/2010 The Islamic Uiversity of Gaza Faculty of Commerce epartmet of Ecoomics ad Political Scieces A Itroductio to Statistics Course (ECOE 30) Sprig Semester 009-00 Fial Eamiatio Solutios 7/6/00 Name: I: Istructor:

More information

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss ECE 90 Lecture : Complexity Regularizatio ad the Squared Loss R. Nowak 5/7/009 I the previous lectures we made use of the Cheroff/Hoeffdig bouds for our aalysis of classifier errors. Hoeffdig s iequality

More information

MODEL CHANGE DETECTION WITH APPLICATION TO MACHINE LEARNING. University of Illinois at Urbana-Champaign

MODEL CHANGE DETECTION WITH APPLICATION TO MACHINE LEARNING. University of Illinois at Urbana-Champaign MODEL CHANGE DETECTION WITH APPLICATION TO MACHINE LEARNING Yuheg Bu Jiaxu Lu Veugopal V. Veeravalli Uiversity of Illiois at Urbaa-Champaig Tsighua Uiversity Email: bu3@illiois.edu, lujx4@mails.tsighua.edu.c,

More information

Mathematical Modeling of Optimum 3 Step Stress Accelerated Life Testing for Generalized Pareto Distribution

Mathematical Modeling of Optimum 3 Step Stress Accelerated Life Testing for Generalized Pareto Distribution America Joural of Theoretical ad Applied Statistics 05; 4(: 6-69 Published olie May 8, 05 (http://www.sciecepublishiggroup.com/j/ajtas doi: 0.648/j.ajtas.05040. ISSN: 6-8999 (Prit; ISSN: 6-9006 (Olie Mathematical

More information

Short Term Load Forecasting Using Artificial Neural Network And Imperialist Competitive Algorithm

Short Term Load Forecasting Using Artificial Neural Network And Imperialist Competitive Algorithm Short Term Load Forecastig Usig Artificial eural etwork Ad Imperialist Competitive Algorithm Mostafa Salamat, Mostafa_salamat63@yahoo.com Javad Mousavi, jmousavi.sh1365@gmail.com Seyed Hamid Shah Alami,

More information

Dimension-free PAC-Bayesian bounds for the estimation of the mean of a random vector

Dimension-free PAC-Bayesian bounds for the estimation of the mean of a random vector Dimesio-free PAC-Bayesia bouds for the estimatio of the mea of a radom vector Olivier Catoi CREST CNRS UMR 9194 Uiversité Paris Saclay olivier.catoi@esae.fr Ilaria Giulii Laboratoire de Probabilités et

More information

The prediction of cement energy demand based on support vector machine Xin Zhang a, *, Shaohong Jing b

The prediction of cement energy demand based on support vector machine Xin Zhang a, *, Shaohong Jing b 4th Iteratioal Coferece o Computer, Mechatroics, Cotrol ad Electroic Egieerig (ICCMCEE 05) The predictio of cemet eergy demad based o support vector machie Xi Zhag a,, Shaohog Jig b School of Electrical

More information

6.883: Online Methods in Machine Learning Alexander Rakhlin

6.883: Online Methods in Machine Learning Alexander Rakhlin 6.883: Olie Methods i Machie Learig Alexader Rakhli LECURE 4 his lecture is partly based o chapters 4-5 i [SSBD4]. Let us o give a variat of SGD for strogly covex fuctios. Algorithm SGD for strogly covex

More information

WHAT IS THE PROBABILITY FUNCTION FOR LARGE TSUNAMI WAVES? ABSTRACT

WHAT IS THE PROBABILITY FUNCTION FOR LARGE TSUNAMI WAVES? ABSTRACT WHAT IS THE PROBABILITY FUNCTION FOR LARGE TSUNAMI WAVES? Harold G. Loomis Hoolulu, HI ABSTRACT Most coastal locatios have few if ay records of tsuami wave heights obtaied over various time periods. Still

More information

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4 MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.

More information

Support Vector Machines and Kernel Methods

Support Vector Machines and Kernel Methods Support Vector Machies ad Kerel Methods Daiel Khashabi Fall 202 Last Update: September 26, 206 Itroductio I Support Vector Machies the goal is to fid a separator betwee data which has the largest margi,

More information

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n. Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator

More information

Random Matrices with Blocks of Intermediate Scale Strongly Correlated Band Matrices

Random Matrices with Blocks of Intermediate Scale Strongly Correlated Band Matrices Radom Matrices with Blocks of Itermediate Scale Strogly Correlated Bad Matrices Jiayi Tog Advisor: Dr. Todd Kemp May 30, 07 Departmet of Mathematics Uiversity of Califoria, Sa Diego Cotets Itroductio Notatio

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization ECE 90 Lecture 4: Maximum Likelihood Estimatio ad Complexity Regularizatio R Nowak 5/7/009 Review : Maximum Likelihood Estimatio We have iid observatios draw from a ukow distributio Y i iid p θ, i,, where

More information

RAINFALL PREDICTION BY WAVELET DECOMPOSITION

RAINFALL PREDICTION BY WAVELET DECOMPOSITION RAIFALL PREDICTIO BY WAVELET DECOMPOSITIO A. W. JAYAWARDEA Departmet of Civil Egieerig, The Uiversit of Hog Kog, Hog Kog, Chia P. C. XU Academ of Mathematics ad Sstem Scieces, Chiese Academ of Scieces,

More information

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ.

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ. 2 5. Weighted umber of late jobs 5.1. Release dates ad due dates: maximimizig the weight of o-time jobs Oce we add release dates, miimizig the umber of late jobs becomes a sigificatly harder problem. For

More information

An Input Domain-Based Reliability Growth Model and Its Applications in Comparing Software Testing Strategies

An Input Domain-Based Reliability Growth Model and Its Applications in Comparing Software Testing Strategies CENTRE NATIONAL DE LA RECHERCHE SCIENTIFIQUE LABORATOIRE D'ANALYSE ET D'ARCHITECTURE DES SYSTÈMES A Iput Domai-Based Reliability Growth Model ad Its Applicatios i Comparig Software Testig Strategies Yiog

More information