On a General Two-Sample Test with New Critical Value for Multivariate Feature Selection in Chest X-Ray Image Analysis Problem
|
|
- Lesley Carroll
- 5 years ago
- Views:
Transcription
1 Applied Mathematical Sciences, Vol. 9, 2015, no. 147, HIKARI Ltd, On a General Two-Sample Test with New Critical Value for Multivariate Feature Selection in Chest X-Ray Image Analysis Problem Samir B. Belhaouari, Hamada R. H. Al-Absi, Ramil F. Kuleev and Nasreddine Megrez Innopolis University, Innopolis, Russia Copyright c 2015 Samir B. Belhaouari, Hamada R. H. Al-Absi, Ramil F. Kuleev and Nasreddine Megrez. This article is distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Abstract In this paper we propose a Two-Sample Test for the means of high dimensional data and a new method to calculate the critical value. The proposed test does not require any condition linking the data dimension and the sample size which makes it a good alternative to the Hotelling T 2 statistic when the data dimension is much larger than the sample size and/or the two sample covariance matrices are not equal. One of the most important application of the proposed test is multivariate feature selection in all fields specially where data dimension is high like image features, genes or finance data. It is also important to highlight the low computing time required by the proposed method to calculate the critical value. Mathematics Subject Classification: 62H15, 60K35, 62G10, 62E20 Keywords: High Dimensional Two Sample Test, Feature Selection 1 Introduction Features selection for classifier is a very important task to reach high accuracy in classification systems. It plays especially important role in complex machine
2 7318 Samir B. Belhaouari et al. learning and computer vision problems as medical image analysis ([14]). Onedimensional metrics measure the overlapping area between classes for a single feature independently of other features. However, in most areas, each feature does not work individually but, rather, with other features, and the correct way to select the best features for classification is to measure the contribution of a set of features together rather than one feature alone. To tackle this task, it is necessary to develop multidimensional metrics. Let { x 1, x 2,..., x n } and { y 1, y 2,..., y m } be two independent random samples generated in an i.i.d. manner from p-dimensional multivariate normal distributions X N( µ 1, Σ 1 ) and Y N( µ 2, Σ 2 ) respectively, where the mean vectors µ 1, µ 2 R p, and the covariance matrices Σ 1, Σ 2 are positive definite. Let us suppose that the mean vectors µ 1, µ 2 and the covariance matrices Σ 1, Σ 2 are unknown. In this paper, we consider the problem of measuring how close are the two Gaussian means vectors µ 1, µ 2 to each other, which translates to testing the high dimensional hypothesis H 0 : µ 1 = µ 2, versus H 1 : µ 1 µ 2. In case X and Y are not Gaussian, it is enough to assume that n and m are sufficiently large, so the Central Limit Theorem applies for samples means. Using maximum likelihood methods, means vectors can be estimated by samples s means, i.e µ 1 = X 1 n = x i, n and µ 2 = Ȳ = 1 m m y i. Hotelling s T 2 test ( [6] ) is the conventional test for the above hypothesis when the dimension p is fixed and less than n + m 2, and Σ 1 = Σ 2. This test is defined as T 2 := nm n + m ( X Ȳ ) T Σ 1 ( X Ȳ ), where Σ is the pooled samples covarince matrix given by Σ = 1 n + m 2 n ( x i X)( xi X) T 1 + n + m 2 = n 1 n + m 2 Σ 1 + m 1 n + m 2 Σ 2. m ( y i Ȳ )( y i Ȳ ) T
3 On a general two-sample test with new critical value 7319 With the Central Limit Theorem, we have n + m (p + 1) p(n + m 2) T 2 F p,n+m (p+1). When p > n + m 2, the matrix Σ is singular and the Hotelling s Test is not defined. As demonstrated in [1], the Hotelling s test is known to be inefficient even when p n + m 2 and is nearly as large as n + m 2. It is also important to highlight the fact that it is hard to verify the assumption Σ 1 = Σ 2 for high demensional data, and using the Hotelling s T 2 test like in [8] may be missleading. Moreover, since the hypothesis H 0 consists of the p-marginal hypotheses: H 0l : µ 1l = µ 2l for l = 1,..., p regarding the means on each data dimension, a natural question is how many hypotheses can be tested simultaneously. These problems were addressed in [4],[7], [3] and [12] with some limitations, complications, and time consuming. Thus, in this paper we provide a simple two sample test that works in all cases without any limitation on p and even when Σ 1 Σ 2. In the context of feature selections, this test estimate the overlapping area between classes. 2 Vector-Variate T statistical metric Let δ := X Ȳ denote the shift vector between the two samples. When H0 holds ( µ 1 = µ 2 ), the mean vector and the covariance matrix of δ are: ( ) ( ) ( ) ( ) E δ = E X Ȳ = E X E Ȳ = µ 1 µ 2 = 0, and [ ] [( ) ( Σ δ = E δ. δ T = E X Ȳ XT Ȳ )] T = Σ X + Σ Ȳ. As the values of the two samples are independant, the covariance matrices Σ X, Σ Ȳ can be evaluated as Σ X = Σ 1 n, and Σ Ȳ = Σ 2 m. Hence, Thus, for n, m big enough, Σ δ = Σ 1 n + Σ 2 m. δ N ( 0, Σ 1 n + Σ ) 2, m
4 7320 Samir B. Belhaouari et al. and we define our Vector-Variate T statistical metric as Z := ( ( ) 1/2 Σ δ Σ1 δ = n + Σ ) 1/2 2 δ N( 0, I), m where I is the identity Matrix. The covariance matrices Σ 1 and Σ 2 are square (p p) and positive definite, so Σ δ is square positive definte and, then, orthogonal diagonalizable. Let Λ : diag( λ 1, λ 2,..., λ p ) and θ := (ϑ 1, ϑ 2,..., ϑ p ) be the matrices formed by the square root of eigenvalues of Σ δ and by corresponding eigenvectors (respectively). Then, Σ δ can be written as Σ δ = (θ 1 Λθ) 2, and our vector variate metric Z can be rewritten as Z := (θ 1 Λ 1 θ) δ. 2.1 The critical value z α The shosen critical value fixes a trade-off between the risk of rejecting H 0 when H 0 actually holds, and the risk of accepting H 0 when H 1 holds. For a significance level α, we reject H 0 in favor of H 1 if Z 2 = z zd 2 > z α, where α = P (z1 2 + z zd 2 > z α). If p = 2 we can take z α = 2 ln(α). For higher dimension p > 2, we define α as: α := 2π 0 = 2π2p 2 2π p π dϕ 0 z α π sin θ 1 dθ 1... sin θ p 2 dθ 2 0 z α r p 1 2π p e r2 2 dr. r p 1 2π p e r2 2 dr Let Φ(x) be the cumulative distribution function of the standard Gaussian variable. We define the sequence with the intial values I(n, x) = x r n e r2 2 dr, I(0, x) = 2π(1 Φ(x)), and I(1, x) = e x2 2. Using integration by parts, we drive the following formula for n > 1: I(n, x) = x n e x π(1 Φ(x)). When n is even, we derive the following recurrence formula I(n, x) = x n 1 e x2 2 + (n 1)I(n 2, x).
5 On a general two-sample test with new critical value Parameters Estimations The covariance matrix needs to be estimated if it is unknown. The unbiased estimator is the sample covariance matrix Σ δ = Σ 1 n + Σ n 2 m = ( x i X)( xi X) T m + ( y i Ȳ )( y i Ȳ ) T. n(n 1) m(m 1) Our estimation needs to be checked and corrected if the covariance estimated has a negative Eigen value. In practice, the negative Eigen values are very close to zero in general, then it is enough to replace all the negative Eigen values by a small positive number noted ε, which it should be smaller, at least 10 times, than the smallest positive Eigen values of the covariance matrix. For example if the first Eigen value is negative, the corrected estimated matrix of covariance is equals to 2 ε Σ δ = θ 1 0 λ θ, λp where θ is the matrix of adapted eigenvectors of the estimated covariance matrix and ε < min k {2,...,p} λ k /10. 3 Simulation 3.1 Feature selection For simulation purpose public JSRT database ([10]) was used. It is the standard digital image database with and without chest lung nodules that was created by the Japanese Society of Radiological Technology (JSRT) in cooperation with the Japanese Radiological Society (JRS) in Database contains 154 nodule and 93 non-nodule images with resolution 2048x2048 pixels. Symlets wavelets at level of decomposition 16 were applied to all images in order to extract features. To reduce the number of features after performing features extraction using Wavelet which results in large number of features, a Multi-feature selection is implemented. The objective is to choose the most significant set of features for classification, i.e. the mixture of features that will best distinguish the object (pathology) from the non-object (normal) classes. To accomplish this, a two-step feature selection process is performed. Proposed statistical metric was used to select best features for classification of regions on image to two classes. Results of our simulation are represented at the Table 1.
6 7322 Samir B. Belhaouari et al. Table 1: Dependency of the accuracy from the number of features Number of features Accuracy False Negative False Positive 1 0,8318 0,2424 0, ,9159 0,1515 0, ,9533 0,0303 0, ,972 0,0303 0, , , , , , , ,9813 0,0303 0, ,9907 0, ,9907 0, Conclusion We have proposed a simple two sample test that works in all cases without any limitation on p and even when Σ 1 Σ 2. This test was particularly conceived for multivariate feature selection like medical image analysis. Due to complexity of the problem each feature does not work indivudually, but tends to work with other features to achieve certain tasks. The Figure 1 shows the inefficiency of one-dimensional metric and the efficiency of two-dimensional metrics: In the context of multivariate feature selection, this test estimates the overlapping area between classes. More precisely, the metric will quantify the contribution of each q features selected among p features, total number of features. The proposed in subsection 2.1 method for calculation of the critical value z α will require reasonable computing time. ( p ) According to the value of the metric, the possible sets of q-features can q be ordered in terms of relevancy for classifier. A method to order the groups of features from the "best group" to the "worst group" is to compare the normalized value of the metric Z, where ν = X + Ȳ. ν 2 Acknowledgements. This work has been supported by the Russian Ministry of education and science with the project "Development of new perspective methods and algorithms for automatic recognition of thorax pathologies based on X-ray images" (agreement: , ID: RFMEF160614X0002) and has been done in Innopolis University.
7 On a general two-sample test with new critical value 7323 Figure 1: Projection on the x-axes or on the y-axes shows a significant overlapping area between the two classes (+ and *), the two features in this example will be described are not significant for classification according to one-dimensional metrics. In other hand, the 2-dimensional metrics will have the capability to well quantify the real overlapping area, i.e. the two features will be described are significant for classification. References [1] Z. Bai, H. Saranadasa, Effect of high dimension: by an example of a two sample problem, Statistica Sinica, 6 (1996), [2] L. Baringhaus, C. Franz, On a new multivariate two-sample test, Journal of Multivariate Analysis, 88 (2004), [3] S. X. Chen, Y. Qin, A two-sample test for high-dimensional data with applications to gene-set testing, The Annals of Statistics, 38 (2010), no. 2, [4] J. Fan, P. Hall and Q. Yao, To how many simultaneous hypothesis tests can normal, Student s t or bootstrap calibration be applied, J. Amer. Statist. Assoc., 102 (2007),
8 7324 Samir B. Belhaouari et al. [5] M. Han, X. Liu, Feature selection techniques with class separability for multivariate time series, Neurocomputing, 110 (2013), [6] H. Hotelling, The generalization of Student s ratio, Annals of Mathematical Statistics, 2 (1931), no. 3, [7] M.R. Kosorok, S. Ma, Marginal asymptotics for the "large p, small n" paradigm: With applications to microarray data, Ann. Statist., 35 (2007), [8] Y. Lu, P.Y. Liu, P. Xiao, and H.W. Deng, Hotelling s T2 multivariate profiling for detecting differential expression in microarrays, Bioinformatics, 21 (2005), no. 14, [9] M. F. Schilling, Multivariate Two-Sample Tests Based on Nearest Neighbors, Journal of the American Statistical Association, 81 (1986), no. 395, [10] J. Shiraishi, S. Katsuragawa, J. Ikezoe, T. Matsumoto, T. Kobayashi, K. Komatsu, M. Matsui, H. Fujita, Y. Kodera and K. Doi, Development of a digital image database for chest radiographs with and without a lung nodule: Receiver operating characteristic analysis of radiologists detection of pulmonary nodules, American Journal of Roentgenology, 174 (2000), no. 1, [11] M. S. Srivastava, S. Katayama, Y. Kano, A two sample test in high dimensional data, Journal of Multivariate Analysis, 114 (2013), [12] M. Van der Laan, J. Bryan, Gene expression analysis with the parametric bootstrap, Biostatistics, 2 (2001), [13] L. Xu, Matrix-Variate Discriminative Analysis, Integrative Hypothesis Testing, and Geno-Pheno A5 Analyzer, Lecture Notes in Computer Science, 7751 (2013), [14] A. N. Zakirov, R. F. Kuleev, A. S. Timoshenko and A. V. Vladimirov, Advanced Approaches to Computer-Aided Detection of Thoracic Diseases on Chest X-Rays, Applied Mathematical Sciences, 9 (2015), no. 88,
9 On a general two-sample test with new critical value 7325 [15] N. Zhou, L. Wang, A Modified T-test Feature Selection Method and Its Application on the HapMap Genotype Data, Genomics, Proteomics and Bioinformatics, 5 (2007), no. 3-4, Received: November 15, 2015; Published: December 17, 2015
Solving Homogeneous Systems with Sub-matrices
Pure Mathematical Sciences, Vol 7, 218, no 1, 11-18 HIKARI Ltd, wwwm-hikaricom https://doiorg/112988/pms218843 Solving Homogeneous Systems with Sub-matrices Massoud Malek Mathematics, California State
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationHigh-dimensional two-sample tests under strongly spiked eigenvalue models
1 High-dimensional two-sample tests under strongly spiked eigenvalue models Makoto Aoshima and Kazuyoshi Yata University of Tsukuba Abstract: We consider a new two-sample test for high-dimensional data
More informationMachine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling
Machine Learning B. Unsupervised Learning B.2 Dimensionality Reduction Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University
More informationSmall sample size in high dimensional space - minimum distance based classification.
Small sample size in high dimensional space - minimum distance based classification. Ewa Skubalska-Rafaj lowicz Institute of Computer Engineering, Automatics and Robotics, Department of Electronics, Wroc
More informationFeature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size
Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size Berkman Sahiner, a) Heang-Ping Chan, Nicholas Petrick, Robert F. Wagner, b) and Lubomir Hadjiiski
More informationStatistical Inference On the High-dimensional Gaussian Covarianc
Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional
More informationEstimation of the Bivariate Generalized. Lomax Distribution Parameters. Based on Censored Samples
Int. J. Contemp. Math. Sciences, Vol. 9, 2014, no. 6, 257-267 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2014.4329 Estimation of the Bivariate Generalized Lomax Distribution Parameters
More informationLecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides
Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Intelligent Data Analysis and Probabilistic Inference Lecture
More informationExplicit Expressions for Free Components of. Sums of the Same Powers
Applied Mathematical Sciences, Vol., 27, no. 53, 2639-2645 HIKARI Ltd, www.m-hikari.com https://doi.org/.2988/ams.27.79276 Explicit Expressions for Free Components of Sums of the Same Powers Alexander
More informationApproximations to the t Distribution
Applied Mathematical Sciences, Vol. 9, 2015, no. 49, 2445-2449 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2015.52148 Approximations to the t Distribution Bashar Zogheib 1 and Ali Elsaheli
More informationSHOTA KATAYAMA AND YUTAKA KANO. Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka , Japan
A New Test on High-Dimensional Mean Vector Without Any Assumption on Population Covariance Matrix SHOTA KATAYAMA AND YUTAKA KANO Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama,
More informationA Practical Method for Decomposition of the Essential Matrix
Applied Mathematical Sciences, Vol. 8, 2014, no. 176, 8755-8770 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.410877 A Practical Method for Decomposition of the Essential Matrix Georgi
More informationResearch Article Sample Size Calculation for Controlling False Discovery Proportion
Probability and Statistics Volume 2012, Article ID 817948, 13 pages doi:10.1155/2012/817948 Research Article Sample Size Calculation for Controlling False Discovery Proportion Shulian Shang, 1 Qianhe Zhou,
More informationECE 661: Homework 10 Fall 2014
ECE 661: Homework 10 Fall 2014 This homework consists of the following two parts: (1) Face recognition with PCA and LDA for dimensionality reduction and the nearest-neighborhood rule for classification;
More informationPoincaré`s Map in a Van der Pol Equation
International Journal of Mathematical Analysis Vol. 8, 014, no. 59, 939-943 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.1988/ijma.014.411338 Poincaré`s Map in a Van der Pol Equation Eduardo-Luis
More informationEmpirical Power of Four Statistical Tests in One Way Layout
International Mathematical Forum, Vol. 9, 2014, no. 28, 1347-1356 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.47128 Empirical Power of Four Statistical Tests in One Way Layout Lorenzo
More informationNovel Approach to Calculation of Box Dimension of Fractal Functions
Applied Mathematical Sciences, vol. 8, 2014, no. 144, 7175-7181 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.49718 Novel Approach to Calculation of Box Dimension of Fractal Functions
More informationA Statistical Analysis of Fukunaga Koontz Transform
1 A Statistical Analysis of Fukunaga Koontz Transform Xiaoming Huo Dr. Xiaoming Huo is an assistant professor at the School of Industrial and System Engineering of the Georgia Institute of Technology,
More informationA Signed-Rank Test Based on the Score Function
Applied Mathematical Sciences, Vol. 10, 2016, no. 51, 2517-2527 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.66189 A Signed-Rank Test Based on the Score Function Hyo-Il Park Department
More information1. Introduction. S.S. Patil 1, Sachidananda 1, U.B. Angadi 2, and D.K. Prabhuraj 3
Cloud Publications International Journal of Advanced Remote Sensing and GIS 2014, Volume 3, Issue 1, pp. 525-531, Article ID Tech-249 ISSN 2320-0243 Research Article Open Access Machine Learning Technique
More informationOn corrections of classical multivariate tests for high-dimensional data
On corrections of classical multivariate tests for high-dimensional data Jian-feng Yao with Zhidong Bai, Dandan Jiang, Shurong Zheng Overview Introduction High-dimensional data and new challenge in statistics
More informationResearch Article Degenerate-Generalized Likelihood Ratio Test for One-Sided Composite Hypotheses
Mathematical Problems in Engineering Volume 2012, Article ID 538342, 11 pages doi:10.1155/2012/538342 Research Article Degenerate-Generalized Likelihood Ratio Test for One-Sided Composite Hypotheses Dongdong
More informationResearch Article A Note on Kantorovich Inequality for Hermite Matrices
Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 0, Article ID 5767, 6 pages doi:0.55/0/5767 Research Article A Note on Kantorovich Inequality for Hermite Matrices Zhibing
More informationConstruction of Combined Charts Based on Combining Functions
Applied Mathematical Sciences, Vol. 8, 2014, no. 84, 4187-4200 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.45359 Construction of Combined Charts Based on Combining Functions Hyo-Il
More informationOn testing the equality of mean vectors in high dimension
ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 17, Number 1, June 2013 Available online at www.math.ut.ee/acta/ On testing the equality of mean vectors in high dimension Muni S.
More informationDouble Gamma Principal Components Analysis
Applied Mathematical Sciences, Vol. 12, 2018, no. 11, 523-533 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2018.8455 Double Gamma Principal Components Analysis Ameerah O. Bahashwan, Zakiah
More informationLECTURE NOTE #3 PROF. ALAN YUILLE
LECTURE NOTE #3 PROF. ALAN YUILLE 1. Three Topics (1) Precision and Recall Curves. Receiver Operating Characteristic Curves (ROC). What to do if we do not fix the loss function? (2) The Curse of Dimensionality.
More informationPermutation-invariant regularization of large covariance matrices. Liza Levina
Liza Levina Permutation-invariant covariance regularization 1/42 Permutation-invariant regularization of large covariance matrices Liza Levina Department of Statistics University of Michigan Joint work
More informationKKM-Type Theorems for Best Proximal Points in Normed Linear Space
International Journal of Mathematical Analysis Vol. 12, 2018, no. 12, 603-609 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ijma.2018.81069 KKM-Type Theorems for Best Proximal Points in Normed
More informationarxiv: v1 [cs.lg] 22 Jun 2009
Bayesian two-sample tests arxiv:0906.4032v1 [cs.lg] 22 Jun 2009 Karsten M. Borgwardt 1 and Zoubin Ghahramani 2 1 Max-Planck-Institutes Tübingen, 2 University of Cambridge June 22, 2009 Abstract In this
More informationEmpirical Comparison of ML and UMVU Estimators of the Generalized Variance for some Normal Stable Tweedie Models: a Simulation Study
Applied Mathematical Sciences, Vol. 10, 2016, no. 63, 3107-3118 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2016.69238 Empirical Comparison of and Estimators of the Generalized Variance for
More informationResearch Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components
Applied Mathematics Volume 202, Article ID 689820, 3 pages doi:0.55/202/689820 Research Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components
More informationMachine Learning. Theory of Classification and Nonparametric Classifier. Lecture 2, January 16, What is theoretically the best classifier
Machine Learning 10-701/15 701/15-781, 781, Spring 2008 Theory of Classification and Nonparametric Classifier Eric Xing Lecture 2, January 16, 2006 Reading: Chap. 2,5 CB and handouts Outline What is theoretically
More informationStatistica Sinica Preprint No: SS R2
Statistica Sinica Preprint No: SS-2016-0063R2 Title Two-sample tests for high-dimension, strongly spiked eigenvalue models Manuscript ID SS-2016-0063R2 URL http://www.stat.sinica.edu.tw/statistica/ DOI
More informationDimensionality Reduction and Principal Components
Dimensionality Reduction and Principal Components Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Motivation Recall, in Bayesian decision theory we have: World: States Y in {1,..., M} and observations of X
More informationOn Powers of General Tridiagonal Matrices
Applied Mathematical Sciences, Vol. 9, 5, no., 583-59 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.988/ams.5.49 On Powers of General Tridiagonal Matrices Qassem M. Al-Hassan Department of Mathematics
More informationGroup Inverse for a Class of. Centrosymmetric Matrix
International athematical Forum, Vol. 13, 018, no. 8, 351-356 HIKARI Ltd, www.m-hikari.com https://doi.org/10.1988/imf.018.8530 Group Inverse for a Class of Centrosymmetric atrix ei Wang and Junqing Wang
More informationWhat is Principal Component Analysis?
What is Principal Component Analysis? Principal component analysis (PCA) Reduce the dimensionality of a data set by finding a new set of variables, smaller than the original set of variables Retains most
More informationNumerical Investigation of the Time Invariant Optimal Control of Singular Systems Using Adomian Decomposition Method
Applied Mathematical Sciences, Vol. 8, 24, no. 2, 6-68 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.2988/ams.24.4863 Numerical Investigation of the Time Invariant Optimal Control of Singular Systems
More informationThe Bayes classifier
The Bayes classifier Consider where is a random vector in is a random variable (depending on ) Let be a classifier with probability of error/risk given by The Bayes classifier (denoted ) is the optimal
More informationIntroduction to Machine Learning
1, DATA11002 Introduction to Machine Learning Lecturer: Teemu Roos TAs: Ville Hyvönen and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer
More informationOn the Laplacian Energy of Windmill Graph. and Graph D m,cn
International Journal of Contemporary Mathematical Sciences Vol. 11, 2016, no. 9, 405-414 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2016.6844 On the Laplacian Energy of Windmill Graph
More informationLearning gradients: prescriptive models
Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan
More informationComputational functional genomics
Computational functional genomics (Spring 2005: Lecture 8) David K. Gifford (Adapted from a lecture by Tommi S. Jaakkola) MIT CSAIL Basic clustering methods hierarchical k means mixture models Multi variate
More informationA Study on Linear and Nonlinear Stiff Problems. Using Single-Term Haar Wavelet Series Technique
Int. Journal of Math. Analysis, Vol. 7, 3, no. 53, 65-636 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.988/ijma.3.3894 A Study on Linear and Nonlinear Stiff Problems Using Single-Term Haar Wavelet Series
More informationMotivating the Covariance Matrix
Motivating the Covariance Matrix Raúl Rojas Computer Science Department Freie Universität Berlin January 2009 Abstract This note reviews some interesting properties of the covariance matrix and its role
More informationExample: Face Detection
Announcements HW1 returned New attendance policy Face Recognition: Dimensionality Reduction On time: 1 point Five minutes or more late: 0.5 points Absent: 0 points Biometrics CSE 190 Lecture 14 CSE190,
More informationBinary Relations in the Space of Binary Relations. I.
Applied Mathematical Sciences, Vol. 8, 2014, no. 109, 5407-5414 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.47515 Binary Relations in the Space of Binary Relations. I. Vyacheslav V.
More informationOn Symmetric Bi-Multipliers of Lattice Implication Algebras
International Mathematical Forum, Vol. 13, 2018, no. 7, 343-350 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/imf.2018.8423 On Symmetric Bi-Multipliers of Lattice Implication Algebras Kyung Ho
More informationSupport Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM
1 Support Vector Machines (SVM) in bioinformatics Day 1: Introduction to SVM Jean-Philippe Vert Bioinformatics Center, Kyoto University, Japan Jean-Philippe.Vert@mines.org Human Genome Center, University
More informationStationary Flows in Acyclic Queuing Networks
Applied Mathematical Sciences, Vol. 11, 2017, no. 1, 23-30 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2017.610257 Stationary Flows in Acyclic Queuing Networks G.Sh. Tsitsiashvili Institute
More informationHopf Bifurcation Analysis of a Dynamical Heart Model with Time Delay
Applied Mathematical Sciences, Vol 11, 2017, no 22, 1089-1095 HIKARI Ltd, wwwm-hikaricom https://doiorg/1012988/ams20177271 Hopf Bifurcation Analysis of a Dynamical Heart Model with Time Delay Luca Guerrini
More informationSome Reviews on Ranks of Upper Triangular Block Matrices over a Skew Field
International Mathematical Forum, Vol 13, 2018, no 7, 323-335 HIKARI Ltd, wwwm-hikaricom https://doiorg/1012988/imf20188528 Some Reviews on Ranks of Upper Triangular lock Matrices over a Skew Field Netsai
More informationPCA and admixture models
PCA and admixture models CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar, Alkes Price PCA and admixture models 1 / 57 Announcements HW1
More informationThe EM Algorithm for the Finite Mixture of Exponential Distribution Models
Int. J. Contemp. Math. Sciences, Vol. 9, 2014, no. 2, 57-64 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijcms.2014.312133 The EM Algorithm for the Finite Mixture of Exponential Distribution
More informationOn Positive Stable Realization for Continuous Linear Singular Systems
Int. Journal of Math. Analysis, Vol. 8, 2014, no. 8, 395-400 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijma.2014.4246 On Positive Stable Realization for Continuous Linear Singular Systems
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Two sample T 2 test 1 Two sample T 2 test 2 Analogous to the univariate context, we
More informationDimension Reduction Techniques. Presented by Jie (Jerry) Yu
Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More informationPCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani
PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given
More informationOn Monitoring Shift in the Mean Processes with. Vector Autoregressive Residual Control Charts of. Individual Observation
Applied Mathematical Sciences, Vol. 8, 14, no. 7, 3491-3499 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.12988/ams.14.44298 On Monitoring Shift in the Mean Processes with Vector Autoregressive Residual
More informationThe Rainbow Connection of Windmill and Corona Graph
Applied Mathematical Sciences, Vol. 8, 2014, no. 128, 6367-6372 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.48632 The Rainbow Connection of Windmill and Corona Graph Yixiao Liu Department
More informationFractal functional regression for classification of gene expression data by wavelets
Fractal functional regression for classification of gene expression data by wavelets Margarita María Rincón 1 and María Dolores Ruiz-Medina 2 1 University of Granada Campus Fuente Nueva 18071 Granada,
More informationEstimation of Stress-Strength Reliability for Kumaraswamy Exponential Distribution Based on Upper Record Values
International Journal of Contemporary Mathematical Sciences Vol. 12, 2017, no. 2, 59-71 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ijcms.2017.7210 Estimation of Stress-Strength Reliability for
More informationINFORMATION THEORY AND STATISTICS
INFORMATION THEORY AND STATISTICS Solomon Kullback DOVER PUBLICATIONS, INC. Mineola, New York Contents 1 DEFINITION OF INFORMATION 1 Introduction 1 2 Definition 3 3 Divergence 6 4 Examples 7 5 Problems...''.
More informationThe Credibility Estimators with Dependence Over Risks
Applied Mathematical Sciences, Vol. 8, 2014, no. 161, 8045-8050 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.410803 The Credibility Estimators with Dependence Over Risks Qiang Zhang
More informationIntroduction to Machine Learning
1, DATA11002 Introduction to Machine Learning Lecturer: Antti Ukkonen TAs: Saska Dönges and Janne Leppä-aho Department of Computer Science University of Helsinki (based in part on material by Patrik Hoyer,
More informationMachine learning for pervasive systems Classification in high-dimensional spaces
Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version
More informationResearch Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance
Advances in Decision Sciences Volume, Article ID 893497, 6 pages doi:.55//893497 Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance Junichi Hirukawa and
More informationLecture 27. December 13, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed
More informationUniversity of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries
University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout :. The Multivariate Gaussian & Decision Boundaries..15.1.5 1 8 6 6 8 1 Mark Gales mjfg@eng.cam.ac.uk Lent
More informationA Study of Relative Efficiency and Robustness of Classification Methods
A Study of Relative Efficiency and Robustness of Classification Methods Yoonkyung Lee* Department of Statistics The Ohio State University *joint work with Rui Wang April 28, 2011 Department of Statistics
More informationHyperbolic Functions and. the Heat Balance Integral Method
Nonl. Analysis and Differential Equations, Vol. 1, 2013, no. 1, 23-27 HIKARI Ltd, www.m-hikari.com Hyperbolic Functions and the Heat Balance Integral Method G. Nhawu and G. Tapedzesa Department of Mathematics,
More informationMethod of Generation of Chaos Map in the Centre Manifold
Advanced Studies in Theoretical Physics Vol. 9, 2015, no. 16, 795-800 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/astp.2015.51097 Method of Generation of Chaos Map in the Centre Manifold Evgeny
More informationSome Properties of a Semi Dynamical System. Generated by von Forester-Losata Type. Partial Equations
Int. Journal of Math. Analysis, Vol. 7, 2013, no. 38, 1863-1868 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijma.2013.3481 Some Properties of a Semi Dynamical System Generated by von Forester-Losata
More informationSufficient Dimension Reduction using Support Vector Machine and it s variants
Sufficient Dimension Reduction using Support Vector Machine and it s variants Andreas Artemiou School of Mathematics, Cardiff University @AG DANK/BCS Meeting 2013 SDR PSVM Real Data Current Research and
More informationMachine Learning Linear Classification. Prof. Matteo Matteucci
Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)
More informationStep-down FDR Procedures for Large Numbers of Hypotheses
Step-down FDR Procedures for Large Numbers of Hypotheses Paul N. Somerville University of Central Florida Abstract. Somerville (2004b) developed FDR step-down procedures which were particularly appropriate
More informationFace Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition
ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr
More informationModel Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model
Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population
More informationA Generalization of p-rings
International Journal of Algebra, Vol. 9, 2015, no. 8, 395-401 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ija.2015.5848 A Generalization of p-rings Adil Yaqub Department of Mathematics University
More informationJackknife Empirical Likelihood Test for Equality of Two High Dimensional Means
Jackknife Empirical Likelihood est for Equality of wo High Dimensional Means Ruodu Wang, Liang Peng and Yongcheng Qi 2 Abstract It has been a long history to test the equality of two multivariate means.
More informationLEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach
LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach Dr. Guangliang Chen February 9, 2016 Outline Introduction Review of linear algebra Matrix SVD PCA Motivation The digits
More informationIntroduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin
1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)
More informationRandom Matrices and Multivariate Statistical Analysis
Random Matrices and Multivariate Statistical Analysis Iain Johnstone, Statistics, Stanford imj@stanford.edu SEA 06@MIT p.1 Agenda Classical multivariate techniques Principal Component Analysis Canonical
More informationA Class of Z4C-Groups
Applied Mathematical Sciences, Vol. 9, 2015, no. 41, 2031-2035 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2015.4121008 A Class of Z4C-Groups Jinshan Zhang 1 School of Science Sichuan University
More informationInterval Images Recognition and Fuzzy Sets
International Mathematical Forum, Vol. 9, 2014, no. 19, 917-921 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.4490 Interval Images Recognition and Fuzzy Sets G. Sh. Tsitsiashvili, Yu.
More informationHakone Seminar Recent Developments in Statistics
Hakone Seminar Recent Developments in Statistics November 12-14, 2015 Hotel Green Plaza Hakone: http://www.hgp.co.jp/language/english/sp/ Organizer: Masanobu TANIGUCHI (Research Institute for Science &
More informationPattern Recognition. Parameter Estimation of Probability Density Functions
Pattern Recognition Parameter Estimation of Probability Density Functions Classification Problem (Review) The classification problem is to assign an arbitrary feature vector x F to one of c classes. The
More informationA review of some semiparametric regression models with application to scoring
A review of some semiparametric regression models with application to scoring Jean-Loïc Berthet 1 and Valentin Patilea 2 1 ENSAI Campus de Ker-Lann Rue Blaise Pascal - BP 37203 35172 Bruz cedex, France
More informationA Generalization of Generalized Triangular Fuzzy Sets
International Journal of Mathematical Analysis Vol, 207, no 9, 433-443 HIKARI Ltd, wwwm-hikaricom https://doiorg/02988/ijma2077350 A Generalization of Generalized Triangular Fuzzy Sets Chang Il Kim Department
More informationThe Automorphisms of a Lie algebra
Applied Mathematical Sciences Vol. 9 25 no. 3 2-27 HIKARI Ltd www.m-hikari.com http://dx.doi.org/.2988/ams.25.4895 The Automorphisms of a Lie algebra WonSok Yoo Department of Applied Mathematics Kumoh
More informationQuadratic Extended Filtering in Nonlinear Systems with Uncertain Observations
Applied Mathematical Sciences, Vol. 8, 2014, no. 4, 157-172 HIKARI Ltd, www.m-hiari.com http://dx.doi.org/10.12988/ams.2014.311636 Quadratic Extended Filtering in Nonlinear Systems with Uncertain Observations
More informationRainbow Connection Number of the Thorn Graph
Applied Mathematical Sciences, Vol. 8, 2014, no. 128, 6373-6377 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.48633 Rainbow Connection Number of the Thorn Graph Yixiao Liu Department
More informationMarginal Screening and Post-Selection Inference
Marginal Screening and Post-Selection Inference Ian McKeague August 13, 2017 Ian McKeague (Columbia University) Marginal Screening August 13, 2017 1 / 29 Outline 1 Background on Marginal Screening 2 2
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction
More informationImprovements in Newton-Rapshon Method for Nonlinear Equations Using Modified Adomian Decomposition Method
International Journal of Mathematical Analysis Vol. 9, 2015, no. 39, 1919-1928 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijma.2015.54124 Improvements in Newton-Rapshon Method for Nonlinear
More informationMachine Learning (BSMC-GA 4439) Wenke Liu
Machine Learning (BSMC-GA 4439) Wenke Liu 02-01-2018 Biomedical data are usually high-dimensional Number of samples (n) is relatively small whereas number of features (p) can be large Sometimes p>>n Problems
More information