On a General Two-Sample Test with New Critical Value for Multivariate Feature Selection in Chest X-Ray Image Analysis Problem

Size: px
Start display at page:

Download "On a General Two-Sample Test with New Critical Value for Multivariate Feature Selection in Chest X-Ray Image Analysis Problem"

Transcription

1 Applied Mathematical Sciences, Vol. 9, 2015, no. 147, HIKARI Ltd, On a General Two-Sample Test with New Critical Value for Multivariate Feature Selection in Chest X-Ray Image Analysis Problem Samir B. Belhaouari, Hamada R. H. Al-Absi, Ramil F. Kuleev and Nasreddine Megrez Innopolis University, Innopolis, Russia Copyright c 2015 Samir B. Belhaouari, Hamada R. H. Al-Absi, Ramil F. Kuleev and Nasreddine Megrez. This article is distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Abstract In this paper we propose a Two-Sample Test for the means of high dimensional data and a new method to calculate the critical value. The proposed test does not require any condition linking the data dimension and the sample size which makes it a good alternative to the Hotelling T 2 statistic when the data dimension is much larger than the sample size and/or the two sample covariance matrices are not equal. One of the most important application of the proposed test is multivariate feature selection in all fields specially where data dimension is high like image features, genes or finance data. It is also important to highlight the low computing time required by the proposed method to calculate the critical value. Mathematics Subject Classification: 62H15, 60K35, 62G10, 62E20 Keywords: High Dimensional Two Sample Test, Feature Selection 1 Introduction Features selection for classifier is a very important task to reach high accuracy in classification systems. It plays especially important role in complex machine

2 7318 Samir B. Belhaouari et al. learning and computer vision problems as medical image analysis ([14]). Onedimensional metrics measure the overlapping area between classes for a single feature independently of other features. However, in most areas, each feature does not work individually but, rather, with other features, and the correct way to select the best features for classification is to measure the contribution of a set of features together rather than one feature alone. To tackle this task, it is necessary to develop multidimensional metrics. Let { x 1, x 2,..., x n } and { y 1, y 2,..., y m } be two independent random samples generated in an i.i.d. manner from p-dimensional multivariate normal distributions X N( µ 1, Σ 1 ) and Y N( µ 2, Σ 2 ) respectively, where the mean vectors µ 1, µ 2 R p, and the covariance matrices Σ 1, Σ 2 are positive definite. Let us suppose that the mean vectors µ 1, µ 2 and the covariance matrices Σ 1, Σ 2 are unknown. In this paper, we consider the problem of measuring how close are the two Gaussian means vectors µ 1, µ 2 to each other, which translates to testing the high dimensional hypothesis H 0 : µ 1 = µ 2, versus H 1 : µ 1 µ 2. In case X and Y are not Gaussian, it is enough to assume that n and m are sufficiently large, so the Central Limit Theorem applies for samples means. Using maximum likelihood methods, means vectors can be estimated by samples s means, i.e µ 1 = X 1 n = x i, n and µ 2 = Ȳ = 1 m m y i. Hotelling s T 2 test ( [6] ) is the conventional test for the above hypothesis when the dimension p is fixed and less than n + m 2, and Σ 1 = Σ 2. This test is defined as T 2 := nm n + m ( X Ȳ ) T Σ 1 ( X Ȳ ), where Σ is the pooled samples covarince matrix given by Σ = 1 n + m 2 n ( x i X)( xi X) T 1 + n + m 2 = n 1 n + m 2 Σ 1 + m 1 n + m 2 Σ 2. m ( y i Ȳ )( y i Ȳ ) T

3 On a general two-sample test with new critical value 7319 With the Central Limit Theorem, we have n + m (p + 1) p(n + m 2) T 2 F p,n+m (p+1). When p > n + m 2, the matrix Σ is singular and the Hotelling s Test is not defined. As demonstrated in [1], the Hotelling s test is known to be inefficient even when p n + m 2 and is nearly as large as n + m 2. It is also important to highlight the fact that it is hard to verify the assumption Σ 1 = Σ 2 for high demensional data, and using the Hotelling s T 2 test like in [8] may be missleading. Moreover, since the hypothesis H 0 consists of the p-marginal hypotheses: H 0l : µ 1l = µ 2l for l = 1,..., p regarding the means on each data dimension, a natural question is how many hypotheses can be tested simultaneously. These problems were addressed in [4],[7], [3] and [12] with some limitations, complications, and time consuming. Thus, in this paper we provide a simple two sample test that works in all cases without any limitation on p and even when Σ 1 Σ 2. In the context of feature selections, this test estimate the overlapping area between classes. 2 Vector-Variate T statistical metric Let δ := X Ȳ denote the shift vector between the two samples. When H0 holds ( µ 1 = µ 2 ), the mean vector and the covariance matrix of δ are: ( ) ( ) ( ) ( ) E δ = E X Ȳ = E X E Ȳ = µ 1 µ 2 = 0, and [ ] [( ) ( Σ δ = E δ. δ T = E X Ȳ XT Ȳ )] T = Σ X + Σ Ȳ. As the values of the two samples are independant, the covariance matrices Σ X, Σ Ȳ can be evaluated as Σ X = Σ 1 n, and Σ Ȳ = Σ 2 m. Hence, Thus, for n, m big enough, Σ δ = Σ 1 n + Σ 2 m. δ N ( 0, Σ 1 n + Σ ) 2, m

4 7320 Samir B. Belhaouari et al. and we define our Vector-Variate T statistical metric as Z := ( ( ) 1/2 Σ δ Σ1 δ = n + Σ ) 1/2 2 δ N( 0, I), m where I is the identity Matrix. The covariance matrices Σ 1 and Σ 2 are square (p p) and positive definite, so Σ δ is square positive definte and, then, orthogonal diagonalizable. Let Λ : diag( λ 1, λ 2,..., λ p ) and θ := (ϑ 1, ϑ 2,..., ϑ p ) be the matrices formed by the square root of eigenvalues of Σ δ and by corresponding eigenvectors (respectively). Then, Σ δ can be written as Σ δ = (θ 1 Λθ) 2, and our vector variate metric Z can be rewritten as Z := (θ 1 Λ 1 θ) δ. 2.1 The critical value z α The shosen critical value fixes a trade-off between the risk of rejecting H 0 when H 0 actually holds, and the risk of accepting H 0 when H 1 holds. For a significance level α, we reject H 0 in favor of H 1 if Z 2 = z zd 2 > z α, where α = P (z1 2 + z zd 2 > z α). If p = 2 we can take z α = 2 ln(α). For higher dimension p > 2, we define α as: α := 2π 0 = 2π2p 2 2π p π dϕ 0 z α π sin θ 1 dθ 1... sin θ p 2 dθ 2 0 z α r p 1 2π p e r2 2 dr. r p 1 2π p e r2 2 dr Let Φ(x) be the cumulative distribution function of the standard Gaussian variable. We define the sequence with the intial values I(n, x) = x r n e r2 2 dr, I(0, x) = 2π(1 Φ(x)), and I(1, x) = e x2 2. Using integration by parts, we drive the following formula for n > 1: I(n, x) = x n e x π(1 Φ(x)). When n is even, we derive the following recurrence formula I(n, x) = x n 1 e x2 2 + (n 1)I(n 2, x).

5 On a general two-sample test with new critical value Parameters Estimations The covariance matrix needs to be estimated if it is unknown. The unbiased estimator is the sample covariance matrix Σ δ = Σ 1 n + Σ n 2 m = ( x i X)( xi X) T m + ( y i Ȳ )( y i Ȳ ) T. n(n 1) m(m 1) Our estimation needs to be checked and corrected if the covariance estimated has a negative Eigen value. In practice, the negative Eigen values are very close to zero in general, then it is enough to replace all the negative Eigen values by a small positive number noted ε, which it should be smaller, at least 10 times, than the smallest positive Eigen values of the covariance matrix. For example if the first Eigen value is negative, the corrected estimated matrix of covariance is equals to 2 ε Σ δ = θ 1 0 λ θ, λp where θ is the matrix of adapted eigenvectors of the estimated covariance matrix and ε < min k {2,...,p} λ k /10. 3 Simulation 3.1 Feature selection For simulation purpose public JSRT database ([10]) was used. It is the standard digital image database with and without chest lung nodules that was created by the Japanese Society of Radiological Technology (JSRT) in cooperation with the Japanese Radiological Society (JRS) in Database contains 154 nodule and 93 non-nodule images with resolution 2048x2048 pixels. Symlets wavelets at level of decomposition 16 were applied to all images in order to extract features. To reduce the number of features after performing features extraction using Wavelet which results in large number of features, a Multi-feature selection is implemented. The objective is to choose the most significant set of features for classification, i.e. the mixture of features that will best distinguish the object (pathology) from the non-object (normal) classes. To accomplish this, a two-step feature selection process is performed. Proposed statistical metric was used to select best features for classification of regions on image to two classes. Results of our simulation are represented at the Table 1.

6 7322 Samir B. Belhaouari et al. Table 1: Dependency of the accuracy from the number of features Number of features Accuracy False Negative False Positive 1 0,8318 0,2424 0, ,9159 0,1515 0, ,9533 0,0303 0, ,972 0,0303 0, , , , , , , ,9813 0,0303 0, ,9907 0, ,9907 0, Conclusion We have proposed a simple two sample test that works in all cases without any limitation on p and even when Σ 1 Σ 2. This test was particularly conceived for multivariate feature selection like medical image analysis. Due to complexity of the problem each feature does not work indivudually, but tends to work with other features to achieve certain tasks. The Figure 1 shows the inefficiency of one-dimensional metric and the efficiency of two-dimensional metrics: In the context of multivariate feature selection, this test estimates the overlapping area between classes. More precisely, the metric will quantify the contribution of each q features selected among p features, total number of features. The proposed in subsection 2.1 method for calculation of the critical value z α will require reasonable computing time. ( p ) According to the value of the metric, the possible sets of q-features can q be ordered in terms of relevancy for classifier. A method to order the groups of features from the "best group" to the "worst group" is to compare the normalized value of the metric Z, where ν = X + Ȳ. ν 2 Acknowledgements. This work has been supported by the Russian Ministry of education and science with the project "Development of new perspective methods and algorithms for automatic recognition of thorax pathologies based on X-ray images" (agreement: , ID: RFMEF160614X0002) and has been done in Innopolis University.

7 On a general two-sample test with new critical value 7323 Figure 1: Projection on the x-axes or on the y-axes shows a significant overlapping area between the two classes (+ and *), the two features in this example will be described are not significant for classification according to one-dimensional metrics. In other hand, the 2-dimensional metrics will have the capability to well quantify the real overlapping area, i.e. the two features will be described are significant for classification. References [1] Z. Bai, H. Saranadasa, Effect of high dimension: by an example of a two sample problem, Statistica Sinica, 6 (1996), [2] L. Baringhaus, C. Franz, On a new multivariate two-sample test, Journal of Multivariate Analysis, 88 (2004), [3] S. X. Chen, Y. Qin, A two-sample test for high-dimensional data with applications to gene-set testing, The Annals of Statistics, 38 (2010), no. 2, [4] J. Fan, P. Hall and Q. Yao, To how many simultaneous hypothesis tests can normal, Student s t or bootstrap calibration be applied, J. Amer. Statist. Assoc., 102 (2007),

8 7324 Samir B. Belhaouari et al. [5] M. Han, X. Liu, Feature selection techniques with class separability for multivariate time series, Neurocomputing, 110 (2013), [6] H. Hotelling, The generalization of Student s ratio, Annals of Mathematical Statistics, 2 (1931), no. 3, [7] M.R. Kosorok, S. Ma, Marginal asymptotics for the "large p, small n" paradigm: With applications to microarray data, Ann. Statist., 35 (2007), [8] Y. Lu, P.Y. Liu, P. Xiao, and H.W. Deng, Hotelling s T2 multivariate profiling for detecting differential expression in microarrays, Bioinformatics, 21 (2005), no. 14, [9] M. F. Schilling, Multivariate Two-Sample Tests Based on Nearest Neighbors, Journal of the American Statistical Association, 81 (1986), no. 395, [10] J. Shiraishi, S. Katsuragawa, J. Ikezoe, T. Matsumoto, T. Kobayashi, K. Komatsu, M. Matsui, H. Fujita, Y. Kodera and K. Doi, Development of a digital image database for chest radiographs with and without a lung nodule: Receiver operating characteristic analysis of radiologists detection of pulmonary nodules, American Journal of Roentgenology, 174 (2000), no. 1, [11] M. S. Srivastava, S. Katayama, Y. Kano, A two sample test in high dimensional data, Journal of Multivariate Analysis, 114 (2013), [12] M. Van der Laan, J. Bryan, Gene expression analysis with the parametric bootstrap, Biostatistics, 2 (2001), [13] L. Xu, Matrix-Variate Discriminative Analysis, Integrative Hypothesis Testing, and Geno-Pheno A5 Analyzer, Lecture Notes in Computer Science, 7751 (2013), [14] A. N. Zakirov, R. F. Kuleev, A. S. Timoshenko and A. V. Vladimirov, Advanced Approaches to Computer-Aided Detection of Thoracic Diseases on Chest X-Rays, Applied Mathematical Sciences, 9 (2015), no. 88,

9 On a general two-sample test with new critical value 7325 [15] N. Zhou, L. Wang, A Modified T-test Feature Selection Method and Its Application on the HapMap Genotype Data, Genomics, Proteomics and Bioinformatics, 5 (2007), no. 3-4, Received: November 15, 2015; Published: December 17, 2015

Solving Homogeneous Systems with Sub-matrices

Solving Homogeneous Systems with Sub-matrices Pure Mathematical Sciences, Vol 7, 218, no 1, 11-18 HIKARI Ltd, wwwm-hikaricom https://doiorg/112988/pms218843 Solving Homogeneous Systems with Sub-matrices Massoud Malek Mathematics, California State

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

High-dimensional two-sample tests under strongly spiked eigenvalue models

High-dimensional two-sample tests under strongly spiked eigenvalue models 1 High-dimensional two-sample tests under strongly spiked eigenvalue models Makoto Aoshima and Kazuyoshi Yata University of Tsukuba Abstract: We consider a new two-sample test for high-dimensional data

More information

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling Machine Learning B. Unsupervised Learning B.2 Dimensionality Reduction Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University

More information

Small sample size in high dimensional space - minimum distance based classification.

Small sample size in high dimensional space - minimum distance based classification. Small sample size in high dimensional space - minimum distance based classification. Ewa Skubalska-Rafaj lowicz Institute of Computer Engineering, Automatics and Robotics, Department of Electronics, Wroc

More information

Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size

Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size Berkman Sahiner, a) Heang-Ping Chan, Nicholas Petrick, Robert F. Wagner, b) and Lubomir Hadjiiski

More information

Statistical Inference On the High-dimensional Gaussian Covarianc

Statistical Inference On the High-dimensional Gaussian Covarianc Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional

More information

Estimation of the Bivariate Generalized. Lomax Distribution Parameters. Based on Censored Samples

Estimation of the Bivariate Generalized. Lomax Distribution Parameters. Based on Censored Samples Int. J. Contemp. Math. Sciences, Vol. 9, 2014, no. 6, 257-267 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2014.4329 Estimation of the Bivariate Generalized Lomax Distribution Parameters

More information

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Intelligent Data Analysis and Probabilistic Inference Lecture

More information

Explicit Expressions for Free Components of. Sums of the Same Powers

Explicit Expressions for Free Components of. Sums of the Same Powers Applied Mathematical Sciences, Vol., 27, no. 53, 2639-2645 HIKARI Ltd, www.m-hikari.com https://doi.org/.2988/ams.27.79276 Explicit Expressions for Free Components of Sums of the Same Powers Alexander

More information

Approximations to the t Distribution

Approximations to the t Distribution Applied Mathematical Sciences, Vol. 9, 2015, no. 49, 2445-2449 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2015.52148 Approximations to the t Distribution Bashar Zogheib 1 and Ali Elsaheli

More information

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan

SHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,

More information

A Practical Method for Decomposition of the Essential Matrix

A Practical Method for Decomposition of the Essential Matrix Applied Mathematical Sciences, Vol. 8, 2014, no. 176, 8755-8770 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.410877 A Practical Method for Decomposition of the Essential Matrix Georgi

More information

Research Article Sample Size Calculation for Controlling False Discovery Proportion

Research Article Sample Size Calculation for Controlling False Discovery Proportion Probability and Statistics Volume 2012, Article ID 817948, 13 pages doi:10.1155/2012/817948 Research Article Sample Size Calculation for Controlling False Discovery Proportion Shulian Shang, 1 Qianhe Zhou,

More information

ECE 661: Homework 10 Fall 2014

ECE 661: Homework 10 Fall 2014 ECE 661: Homework 10 Fall 2014 This homework consists of the following two parts: (1) Face recognition with PCA and LDA for dimensionality reduction and the nearest-neighborhood rule for classification;

More information

Poincaré`s Map in a Van der Pol Equation

Poincaré`s Map in a Van der Pol Equation International Journal of Mathematical Analysis Vol. 8, 014, no. 59, 939-943 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.1988/ijma.014.411338 Poincaré`s Map in a Van der Pol Equation Eduardo-Luis

More information

Empirical Power of Four Statistical Tests in One Way Layout

Empirical Power of Four Statistical Tests in One Way Layout International Mathematical Forum, Vol. 9, 2014, no. 28, 1347-1356 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.47128 Empirical Power of Four Statistical Tests in One Way Layout Lorenzo

More information

Novel Approach to Calculation of Box Dimension of Fractal Functions

Novel Approach to Calculation of Box Dimension of Fractal Functions Applied Mathematical Sciences, vol. 8, 2014, no. 144, 7175-7181 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.49718 Novel Approach to Calculation of Box Dimension of Fractal Functions

More information

A Statistical Analysis of Fukunaga Koontz Transform

A Statistical Analysis of Fukunaga Koontz Transform 1 A Statistical Analysis of Fukunaga Koontz Transform Xiaoming Huo Dr. Xiaoming Huo is an assistant professor at the School of Industrial and System Engineering of the Georgia Institute of Technology,

More information

A Signed-Rank Test Based on the Score Function

A Signed-Rank Test Based on the Score Function Applied Mathematical Sciences, Vol. 10, 2016, no. 51, 2517-2527 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.66189 A Signed-Rank Test Based on the Score Function Hyo-Il Park Department

More information

1. Introduction. S.S. Patil 1, Sachidananda 1, U.B. Angadi 2, and D.K. Prabhuraj 3

1. Introduction. S.S. Patil 1, Sachidananda 1, U.B. Angadi 2, and D.K. Prabhuraj 3 Cloud Publications International Journal of Advanced Remote Sensing and GIS 2014, Volume 3, Issue 1, pp. 525-531, Article ID Tech-249 ISSN 2320-0243 Research Article Open Access Machine Learning Technique

More information

On corrections of classical multivariate tests for high-dimensional data

On corrections of classical multivariate tests for high-dimensional data On corrections of classical multivariate tests for high-dimensional data Jian-feng Yao with Zhidong Bai, Dandan Jiang, Shurong Zheng Overview Introduction High-dimensional data and new challenge in statistics

More information

Research Article Degenerate-Generalized Likelihood Ratio Test for One-Sided Composite Hypotheses

Research Article Degenerate-Generalized Likelihood Ratio Test for One-Sided Composite Hypotheses Mathematical Problems in Engineering Volume 2012, Article ID 538342, 11 pages doi:10.1155/2012/538342 Research Article Degenerate-Generalized Likelihood Ratio Test for One-Sided Composite Hypotheses Dongdong

More information

Research Article A Note on Kantorovich Inequality for Hermite Matrices

Research Article A Note on Kantorovich Inequality for Hermite Matrices Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 0, Article ID 5767, 6 pages doi:0.55/0/5767 Research Article A Note on Kantorovich Inequality for Hermite Matrices Zhibing

More information

Construction of Combined Charts Based on Combining Functions

Construction of Combined Charts Based on Combining Functions Applied Mathematical Sciences, Vol. 8, 2014, no. 84, 4187-4200 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.45359 Construction of Combined Charts Based on Combining Functions Hyo-Il

More information

On testing the equality of mean vectors in high dimension

On testing the equality of mean vectors in high dimension ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 17, Number 1, June 2013 Available online at www.math.ut.ee/acta/ On testing the equality of mean vectors in high dimension Muni S.

More information

Double Gamma Principal Components Analysis

Double Gamma Principal Components Analysis Applied Mathematical Sciences, Vol. 12, 2018, no. 11, 523-533 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2018.8455 Double Gamma Principal Components Analysis Ameerah O. Bahashwan, Zakiah

More information

LECTURE NOTE #3 PROF. ALAN YUILLE

LECTURE NOTE #3 PROF. ALAN YUILLE LECTURE NOTE #3 PROF. ALAN YUILLE 1. Three Topics (1) Precision and Recall Curves. Receiver Operating Characteristic Curves (ROC). What to do if we do not fix the loss function? (2) The Curse of Dimensionality.

More information

Permutation-invariant regularization of large covariance matrices. Liza Levina

Permutation-invariant regularization of large covariance matrices. Liza Levina Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work

More information

KKM-Type Theorems for Best Proximal Points in Normed Linear Space

KKM-Type Theorems for Best Proximal Points in Normed Linear Space International Journal of Mathematical Analysis Vol. 12, 2018, no. 12, 603-609 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ijma.2018.81069 KKM-Type Theorems for Best Proximal Points in Normed

More information

arxiv: v1 [cs.lg] 22 Jun 2009

arxiv: v1 [cs.lg] 22 Jun 2009 Bayesian two-sample tests arxiv:0906.4032v1 [cs.lg] 22 Jun 2009 Karsten M. Borgwardt 1 and Zoubin Ghahramani 2 1 Max-Planck-Institutes Tübingen, 2 University of Cambridge June 22, 2009 Abstract In this

More information

Empirical Comparison of ML and UMVU Estimators of the Generalized Variance for some Normal Stable Tweedie Models: a Simulation Study

Empirical Comparison of ML and UMVU Estimators of the Generalized Variance for some Normal Stable Tweedie Models: a Simulation Study Applied Mathematical Sciences, Vol. 10, 2016, no. 63, 3107-3118 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2016.69238 Empirical Comparison of and Estimators of the Generalized Variance for

More information

Research Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components

Research Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components Applied Mathematics Volume 202, Article ID 689820, 3 pages doi:0.55/202/689820 Research Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components

More information

Machine Learning. Theory of Classification and Nonparametric Classifier. Lecture 2, January 16, What is theoretically the best classifier

Machine Learning. Theory of Classification and Nonparametric Classifier. Lecture 2, January 16, What is theoretically the best classifier Machine Learning 10-701/15 701/15-781, 781, Spring 2008 Theory of Classification and Nonparametric Classifier Eric Xing Lecture 2, January 16, 2006 Reading: Chap. 2,5 CB and handouts Outline What is theoretically

More information

Statistica Sinica Preprint No: SS R2

Statistica Sinica Preprint No: SS R2 Statistica Sinica Preprint No: SS-2016-0063R2 Title Two-sample tests for high-dimension, strongly spiked eigenvalue models Manuscript ID SS-2016-0063R2 URL http://www.stat.sinica.edu.tw/statistica/ DOI

More information

Dimensionality Reduction and Principal Components

Dimensionality Reduction and Principal Components Dimensionality Reduction and Principal Components Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Motivation Recall, in Bayesian decision theory we have: World: States Y in {1,..., M} and observations of X

More information

On Powers of General Tridiagonal Matrices

On Powers of General Tridiagonal Matrices Applied Mathematical Sciences, Vol. 9, 5, no., 583-59 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.988/ams.5.49 On Powers of General Tridiagonal Matrices Qassem M. Al-Hassan Department of Mathematics

More information

Group Inverse for a Class of. Centrosymmetric Matrix

Group Inverse for a Class of. Centrosymmetric Matrix International athematical Forum, Vol. 13, 018, no. 8, 351-356 HIKARI Ltd, www.m-hikari.com https://doi.org/10.1988/imf.018.8530 Group Inverse for a Class of Centrosymmetric atrix ei Wang and Junqing Wang

More information

What is Principal Component Analysis?

What is Principal Component Analysis? What is Principal Component Analysis? Principal component analysis (PCA) Reduce the dimensionality of a data set by finding a new set of variables, smaller than the original set of variables Retains most

More information

Numerical Investigation of the Time Invariant Optimal Control of Singular Systems Using Adomian Decomposition Method

Numerical Investigation of the Time Invariant Optimal Control of Singular Systems Using Adomian Decomposition Method Applied Mathematical Sciences, Vol. 8, 24, no. 2, 6-68 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.2988/ams.24.4863 Numerical Investigation of the Time Invariant Optimal Control of Singular Systems

More information

The Bayes classifier

The Bayes classifier The Bayes classifier Consider where is a random vector in is a random variable (depending on ) Let be a classifier with probability of error/risk given by The Bayes classifier (denoted ) is the optimal

More information

Introduction to Machine Learning

Introduction to Machine Learning 1, DATA11002 Introduction to Machine Learning Lecturer: Teemu Roos TAs: Ville Hyvönen and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer

More information

On the Laplacian Energy of Windmill Graph. and Graph D m,cn

On the Laplacian Energy of Windmill Graph. and Graph D m,cn International Journal of Contemporary Mathematical Sciences Vol. 11, 2016, no. 9, 405-414 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2016.6844 On the Laplacian Energy of Windmill Graph

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Computational functional genomics

Computational functional genomics Computational functional genomics (Spring 2005: Lecture 8) David K. Gifford (Adapted from a lecture by Tommi S. Jaakkola) MIT CSAIL Basic clustering methods hierarchical k means mixture models Multi variate

More information

A Study on Linear and Nonlinear Stiff Problems. Using Single-Term Haar Wavelet Series Technique

A Study on Linear and Nonlinear Stiff Problems. Using Single-Term Haar Wavelet Series Technique Int. Journal of Math. Analysis, Vol. 7, 3, no. 53, 65-636 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.988/ijma.3.3894 A Study on Linear and Nonlinear Stiff Problems Using Single-Term Haar Wavelet Series

More information

Motivating the Covariance Matrix

Motivating the Covariance Matrix Motivating the Covariance Matrix Raúl Rojas Computer Science Department Freie Universität Berlin January 2009 Abstract This note reviews some interesting properties of the covariance matrix and its role

More information

Example: Face Detection

Example: Face Detection Announcements HW1 returned New attendance policy Face Recognition: Dimensionality Reduction On time: 1 point Five minutes or more late: 0.5 points Absent: 0 points Biometrics CSE 190 Lecture 14 CSE190,

More information

Binary Relations in the Space of Binary Relations. I.

Binary Relations in the Space of Binary Relations. I. Applied Mathematical Sciences, Vol. 8, 2014, no. 109, 5407-5414 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.47515 Binary Relations in the Space of Binary Relations. I. Vyacheslav V.

More information

On Symmetric Bi-Multipliers of Lattice Implication Algebras

On Symmetric Bi-Multipliers of Lattice Implication Algebras International Mathematical Forum, Vol. 13, 2018, no. 7, 343-350 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/imf.2018.8423 On Symmetric Bi-Multipliers of Lattice Implication Algebras Kyung Ho

More information

Support Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM

Support Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM 1 Support Vector Machines (SVM) in bioinformatics Day 1: Introduction to SVM Jean-Philippe Vert Bioinformatics Center, Kyoto University, Japan Jean-Philippe.Vert@mines.org Human Genome Center, University

More information

Stationary Flows in Acyclic Queuing Networks

Stationary Flows in Acyclic Queuing Networks Applied Mathematical Sciences, Vol. 11, 2017, no. 1, 23-30 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2017.610257 Stationary Flows in Acyclic Queuing Networks G.Sh. Tsitsiashvili Institute

More information

Hopf Bifurcation Analysis of a Dynamical Heart Model with Time Delay

Hopf Bifurcation Analysis of a Dynamical Heart Model with Time Delay Applied Mathematical Sciences, Vol 11, 2017, no 22, 1089-1095 HIKARI Ltd, wwwm-hikaricom https://doiorg/1012988/ams20177271 Hopf Bifurcation Analysis of a Dynamical Heart Model with Time Delay Luca Guerrini

More information

Some Reviews on Ranks of Upper Triangular Block Matrices over a Skew Field

Some Reviews on Ranks of Upper Triangular Block Matrices over a Skew Field International Mathematical Forum, Vol 13, 2018, no 7, 323-335 HIKARI Ltd, wwwm-hikaricom https://doiorg/1012988/imf20188528 Some Reviews on Ranks of Upper Triangular lock Matrices over a Skew Field Netsai

More information

PCA and admixture models

PCA and admixture models PCA and admixture models CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar, Alkes Price PCA and admixture models 1 / 57 Announcements HW1

More information

The EM Algorithm for the Finite Mixture of Exponential Distribution Models

The EM Algorithm for the Finite Mixture of Exponential Distribution Models Int. J. Contemp. Math. Sciences, Vol. 9, 2014, no. 2, 57-64 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2014.312133 The EM Algorithm for the Finite Mixture of Exponential Distribution

More information

On Positive Stable Realization for Continuous Linear Singular Systems

On Positive Stable Realization for Continuous Linear Singular Systems Int. Journal of Math. Analysis, Vol. 8, 2014, no. 8, 395-400 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijma.2014.4246 On Positive Stable Realization for Continuous Linear Singular Systems

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Two sample T 2 test 1 Two sample T 2 test 2 Analogous to the univariate context, we

More information

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given

More information

On Monitoring Shift in the Mean Processes with. Vector Autoregressive Residual Control Charts of. Individual Observation

On Monitoring Shift in the Mean Processes with. Vector Autoregressive Residual Control Charts of. Individual Observation Applied Mathematical Sciences, Vol. 8, 14, no. 7, 3491-3499 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.12988/ams.14.44298 On Monitoring Shift in the Mean Processes with Vector Autoregressive Residual

More information

The Rainbow Connection of Windmill and Corona Graph

The Rainbow Connection of Windmill and Corona Graph Applied Mathematical Sciences, Vol. 8, 2014, no. 128, 6367-6372 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.48632 The Rainbow Connection of Windmill and Corona Graph Yixiao Liu Department

More information

Fractal functional regression for classification of gene expression data by wavelets

Fractal functional regression for classification of gene expression data by wavelets Fractal functional regression for classification of gene expression data by wavelets Margarita María Rincón 1 and María Dolores Ruiz-Medina 2 1 University of Granada Campus Fuente Nueva 18071 Granada,

More information

Estimation of Stress-Strength Reliability for Kumaraswamy Exponential Distribution Based on Upper Record Values

Estimation of Stress-Strength Reliability for Kumaraswamy Exponential Distribution Based on Upper Record Values International Journal of Contemporary Mathematical Sciences Vol. 12, 2017, no. 2, 59-71 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ijcms.2017.7210 Estimation of Stress-Strength Reliability for

More information

INFORMATION THEORY AND STATISTICS

INFORMATION THEORY AND STATISTICS INFORMATION THEORY AND STATISTICS Solomon Kullback DOVER PUBLICATIONS, INC. Mineola, New York Contents 1 DEFINITION OF INFORMATION 1 Introduction 1 2 Definition 3 3 Divergence 6 4 Examples 7 5 Problems...''.

More information

The Credibility Estimators with Dependence Over Risks

The Credibility Estimators with Dependence Over Risks Applied Mathematical Sciences, Vol. 8, 2014, no. 161, 8045-8050 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.410803 The Credibility Estimators with Dependence Over Risks Qiang Zhang

More information

Introduction to Machine Learning

Introduction to Machine Learning 1, DATA11002 Introduction to Machine Learning Lecturer: Antti Ukkonen TAs: Saska Dönges and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer,

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance

Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance Advances in Decision Sciences Volume, Article ID 893497, 6 pages doi:.55//893497 Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance Junichi Hirukawa and

More information

Lecture 27. December 13, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Lecture 27. December 13, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University. This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed

More information

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout :. The Multivariate Gaussian & Decision Boundaries..15.1.5 1 8 6 6 8 1 Mark Gales mjfg@eng.cam.ac.uk Lent

More information

A Study of Relative Efficiency and Robustness of Classification Methods

A Study of Relative Efficiency and Robustness of Classification Methods A Study of Relative Efficiency and Robustness of Classification Methods Yoonkyung Lee* Department of Statistics The Ohio State University *joint work with Rui Wang April 28, 2011 Department of Statistics

More information

Hyperbolic Functions and. the Heat Balance Integral Method

Hyperbolic Functions and. the Heat Balance Integral Method Nonl. Analysis and Differential Equations, Vol. 1, 2013, no. 1, 23-27 HIKARI Ltd, www.m-hikari.com Hyperbolic Functions and the Heat Balance Integral Method G. Nhawu and G. Tapedzesa Department of Mathematics,

More information

Method of Generation of Chaos Map in the Centre Manifold

Method of Generation of Chaos Map in the Centre Manifold Advanced Studies in Theoretical Physics Vol. 9, 2015, no. 16, 795-800 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/astp.2015.51097 Method of Generation of Chaos Map in the Centre Manifold Evgeny

More information

Some Properties of a Semi Dynamical System. Generated by von Forester-Losata Type. Partial Equations

Some Properties of a Semi Dynamical System. Generated by von Forester-Losata Type. Partial Equations Int. Journal of Math. Analysis, Vol. 7, 2013, no. 38, 1863-1868 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijma.2013.3481 Some Properties of a Semi Dynamical System Generated by von Forester-Losata

More information

Sufficient Dimension Reduction using Support Vector Machine and it s variants

Sufficient Dimension Reduction using Support Vector Machine and it s variants Sufficient Dimension Reduction using Support Vector Machine and it s variants Andreas Artemiou School of Mathematics, Cardiff University @AG DANK/BCS Meeting 2013 SDR PSVM Real Data Current Research and

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

Step-down FDR Procedures for Large Numbers of Hypotheses

Step-down FDR Procedures for Large Numbers of Hypotheses Step-down FDR Procedures for Large Numbers of Hypotheses Paul N. Somerville University of Central Florida Abstract. Somerville (2004b) developed FDR step-down procedures which were particularly appropriate

More information

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr

More information

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population

More information

A Generalization of p-rings

A Generalization of p-rings International Journal of Algebra, Vol. 9, 2015, no. 8, 395-401 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ija.2015.5848 A Generalization of p-rings Adil Yaqub Department of Mathematics University

More information

Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means

Jackknife Empirical Likelihood Test for Equality of Two High Dimensional Means Jackknife Empirical Likelihood est for Equality of wo High Dimensional Means Ruodu Wang, Liang Peng and Yongcheng Qi 2 Abstract It has been a long history to test the equality of two multivariate means.

More information

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach Dr. Guangliang Chen February 9, 2016 Outline Introduction Review of linear algebra Matrix SVD PCA Motivation The digits

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Random Matrices and Multivariate Statistical Analysis

Random Matrices and Multivariate Statistical Analysis Random Matrices and Multivariate Statistical Analysis Iain Johnstone, Statistics, Stanford imj@stanford.edu SEA 06@MIT p.1 Agenda Classical multivariate techniques Principal Component Analysis Canonical

More information

A Class of Z4C-Groups

A Class of Z4C-Groups Applied Mathematical Sciences, Vol. 9, 2015, no. 41, 2031-2035 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2015.4121008 A Class of Z4C-Groups Jinshan Zhang 1 School of Science Sichuan University

More information

Interval Images Recognition and Fuzzy Sets

Interval Images Recognition and Fuzzy Sets International Mathematical Forum, Vol. 9, 2014, no. 19, 917-921 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.4490 Interval Images Recognition and Fuzzy Sets G. Sh. Tsitsiashvili, Yu.

More information

Hakone Seminar Recent Developments in Statistics

Hakone Seminar Recent Developments in Statistics Hakone Seminar Recent Developments in Statistics November 12-14, 2015 Hotel Green Plaza Hakone: http://www.hgp.co.jp/language/english/sp/ Organizer: Masanobu TANIGUCHI (Research Institute for Science &

More information

Pattern Recognition. Parameter Estimation of Probability Density Functions

Pattern Recognition. Parameter Estimation of Probability Density Functions Pattern Recognition Parameter Estimation of Probability Density Functions Classification Problem (Review) The classification problem is to assign an arbitrary feature vector x F to one of c classes. The

More information

A review of some semiparametric regression models with application to scoring

A review of some semiparametric regression models with application to scoring A review of some semiparametric regression models with application to scoring Jean-Loïc Berthet 1 and Valentin Patilea 2 1 ENSAI Campus de Ker-Lann Rue Blaise Pascal - BP 37203 35172 Bruz cedex, France

More information

A Generalization of Generalized Triangular Fuzzy Sets

A Generalization of Generalized Triangular Fuzzy Sets International Journal of Mathematical Analysis Vol, 207, no 9, 433-443 HIKARI Ltd, wwwm-hikaricom https://doiorg/02988/ijma2077350 A Generalization of Generalized Triangular Fuzzy Sets Chang Il Kim Department

More information

The Automorphisms of a Lie algebra

The Automorphisms of a Lie algebra Applied Mathematical Sciences Vol. 9 25 no. 3 2-27 HIKARI Ltd www.m-hikari.com http://dx.doi.org/.2988/ams.25.4895 The Automorphisms of a Lie algebra WonSok Yoo Department of Applied Mathematics Kumoh

More information

Quadratic Extended Filtering in Nonlinear Systems with Uncertain Observations

Quadratic Extended Filtering in Nonlinear Systems with Uncertain Observations Applied Mathematical Sciences, Vol. 8, 2014, no. 4, 157-172 HIKARI Ltd, www.m-hiari.com http://dx.doi.org/10.12988/ams.2014.311636 Quadratic Extended Filtering in Nonlinear Systems with Uncertain Observations

More information

Rainbow Connection Number of the Thorn Graph

Rainbow Connection Number of the Thorn Graph Applied Mathematical Sciences, Vol. 8, 2014, no. 128, 6373-6377 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.48633 Rainbow Connection Number of the Thorn Graph Yixiao Liu Department

More information

Marginal Screening and Post-Selection Inference

Marginal Screening and Post-Selection Inference Marginal Screening and Post-Selection Inference Ian McKeague August 13, 2017 Ian McKeague (Columbia University) Marginal Screening August 13, 2017 1 / 29 Outline 1 Background on Marginal Screening 2 2

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction

More information

Improvements in Newton-Rapshon Method for Nonlinear Equations Using Modified Adomian Decomposition Method

Improvements in Newton-Rapshon Method for Nonlinear Equations Using Modified Adomian Decomposition Method International Journal of Mathematical Analysis Vol. 9, 2015, no. 39, 1919-1928 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijma.2015.54124 Improvements in Newton-Rapshon Method for Nonlinear

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 02-01-2018 Biomedical data are usually high-dimensional Number of samples (n) is relatively small whereas number of features (p) can be large Sometimes p>>n Problems

More information