Computation of Information Value for Credit Scoring Models
|
|
- Austin Moore
- 5 years ago
- Views:
Transcription
1 Jedovnice 20 Computation of Information Value for Credit Scoring Models Martin Řezáč, Jan Koláček Dept. of Mathematics and Statistics, Faculty of Science, Masaryk University
2 Information value The special case of Kullback-Leibler divergence given by: where are densities of scores of bad and good clients. Finanční matematika v praxi I, Jedovnice 20 /30
3 Kernel estimate The kernel density estimates are defined by where and K is the Epanechnikov kernel. Finanční matematika v praxi I, Jedovnice 20 2/30
4 Empirical estimate of I val Finanční matematika v praxi I, Jedovnice 20 3/30
5 Empirical estimate of I val However in practice, there could occur computational problems. The Information value index becomes infinite in cases when some of n 0 j or n j are equal to 0. Choosing of the number of bins is also very important. In the literature and also in many applications in credit scoring, the value r=0 is preferred. Finanční matematika v praxi I, Jedovnice 20 4/30
6 Empirical estimate with supervised interval selection (ESIS), We want to avoid zero values of n 0 j or n j. We propose to require to have at least k, where k is a positive integer, observations of scores of both good and bad client in each interval. This is the basic idea of all proposed algorithms. Finanční matematika v praxi I, Jedovnice 20 5/30
7 Empirical estimate with supervised interval selection, the first ESIS: Set where is the empirical quantile function appropriate to the empirical cumulative distribution function of scores of bad clients. Finanční matematika v praxi I, Jedovnice 20 6/30
8 Empirical estimate with supervised interval selection Usage of quantile function of scores of bad clients is motivated by the assumption, that number of bad clients is less than number of good clients. If n 0 is not divisible by k, it is necessary to adjust our intervals, because we obtain number of scores of bad clients in the last interval, which is less than k. In this case, we have to merge the last two intervals. Furthermore we need to ensure, that the number of scores of good clients is as required in each interval. To do so, we compute n j for all actual intervals. If we obtain n j < k for jth interval, we merge this interval with its neighbor on the right side. This can be done for all intervals except the last one. If we have n < k for the j last interval, than we have to merge it with its neighbor on the left side, i.e. we merge the last two intervals. Finanční matematika v praxi I, Jedovnice 20 7/30
9 and Empirical estimate with supervised interval selection Very important is the choice of k. If we choose too small value, we get overestimated value of the Information value, and vice versa. As a reasonable compromise seems to be adjusted square root of number of bad clients given by The estimate of the Information value is given by where n and n 0 correspond to observed counts of good and bad clients j j in intervals created according to the described procedure. Finanční matematika v praxi I, Jedovnice 20 8/30
10 Simulation results Consider n clients, 00p B % of bad clients with and 00(-p B )% of good clients with f N(, ). : 0 Because of normality we know I val. Consider following values of parameters: n = , n = 000 μ 0 = 0 σ 0 = σ = μ = 0.5,,.5 p B = 0.02, 0.05, 0., f : N( 0, 0) 0 Finanční matematika v praxi I, Jedovnice 20 9/30
11 Simulation results ) Scores of bad and good clients were generated according to given parameters. 2) Estimates ˆ, Iˆ, Iˆ were computed. 3) Square errors were computed. 4) Steps )-3) were repeated one thousand times. 5) was computed. I val, DEC val, KERN val, ESIS Finanční matematika v praxi I, Jedovnice 20 0/30
12 Simulation results n=00000, = IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 n=00000, =.0 0 IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , n=000, = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , n=000, =.0 0 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , n=000, =.5 0 IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 worst average best performance Finanční matematika v praxi I, Jedovnice 20 /30
13 and Adjusted empirical estimate with supervised interval selection (AESIS) It is obvious that the choice of parameter k is crucial. So the question is: Is the choice optimal (according to )? What effect has n 0 on the optimal k? And what effect, if any, has the difference of means? 0 Finanční matematika v praxi I, Jedovnice 20 2/30
14 Simulation results Consider 0000 clients, 00p B % of bad clients with 00(-p B )% of good clients with f : N(,). Set 0 0 and consider 0.5,and.5,. E(( Iˆ val I val ) 2 ) f : N( 0,) 0 and k p B p B 0 0 3/30
15 Simulation results Dependence of on k,. 0 The highlighted circles correspond to values of k, where minimal value of the is obtained. The diamonds correspond to values of k given by. 0 p B k 0 p B ˆ I val, AESIS 4/30
16 Simulation results n=000, = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, , , , n=000, =.0 0 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_aesis 0,2568 0, , , n=000, =.5 0 The classical estimate had the lowest in case of weak model and extremely low number (namely 20) of bad clients. If the number of bad client is slightly higher, had the best performance. Considering models with higher predictive power, and had the best performance. IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 IV_aesis 0, , ,3587 0, Finanční matematika v praxi I, Jedovnice 20 5/30
17 Simulation results n=00000, = IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 IV_aesis 0, , , , n=00000, =.0 0 IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, ,0057 0, ,0003 n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , IV_aesis 0, , , ,0472 We can see that the classical estimate was outperformed by all other estimates of the Information value in all considered cases when the number of clients was high (00000 in our case). The best performance was achieved by in case of weak scoring models, by in case of models with high performance and by for models with very high performance. Finanční matematika v praxi I, Jedovnice 20 5/30
18 Algorithm for the modified ESIS: ) 2) 3) q [] k q j F n smax max( q j, q j 0 ) q j 0 F 0 k n ESIS. 0, ˆ I val, ESIS where s max 4) Add to the sequence, i.e. q [ q, smax ] 5) Erase all scores s max 6) While n 0 and n are greater than 2*k, repeat step 2) 5) 7) q [min( score ), q] Finanční matematika v praxi I, Jedovnice 20 7/30
19 and AESIS. Simulation results Consider 000 and 0000 clients, 00p B % of bad clients with and 00(-p B )% of good clients with f : N(,). Set 0 0 and consider 0.5,and.5,. E(( Iˆ val I val ) 2 ) f : N( 0,) 0 n 000 n 0000 k p B k pb Finanční matematika v praxi I, Jedovnice 20 8/30
20 Simulation results Dependence of on k. n 000, p 0.2 B n 0000, p 0.2 B k pb n 3. 4 ˆ ˆ0 n 000 pb n k 3. 4 ˆ ˆ0 0 n 0000 p B k 0 p B ˆ I val, AESIS pb n k 3. 4 ˆ ˆ0 0 p B k 0 p B /30
21 Simulation results n=00000, = IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 IV_aesis 0, , , , IV_esis 0, , , ,00079 IV_aesis 0, , , ,00023 n=00000, =.0 0 IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, ,0057 0, ,0003 IV_esis 0, , , ,034 IV_aesis 0, , , ,00534 n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , IV_aesis 0, , , ,0472 IV_esis 0, , ,297 0,6950 IV_aesis 0, , , , Finanční matematika v praxi I, Jedovnice 20 n=000, = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, , , , IV_esis 0,0279 0, , , IV_aesis 0, , , ,00903 n=000, =.0 0 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_aesis 0,2568 0, , , IV_esis 0, ,4470 0,5949 0, IV_aesis 0, , , , n=000, =.5 0 IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 IV_aesis 0, , ,3587 0, IV_esis,50276, , ,86092 IV_aesis 0, , ,2057 0, /30
22 . Simulation results The new algorithm (ESIS.) ended disastrously in most cases. The only exception was a situation where n=000 a μ -μ 0 =0.5. This corresponds to a scoring model with very poor discriminatory power and a very small number of observed bad clients (20-200). AESIS. ranked among the average. Finanční matematika v praxi I, Jedovnice 20 2/30
23 ESIS.2 U původního ESIS často dochází ke slučování vypočtených intervalů ve druhé fázi algoritmu. Pro výpočet se používá jen F0.. Aby byla splněna podmínka n >k, je zřejmě nutné, aby hranice prvního intervalu byla větší než k. F n To vede k myšlence použít ke konstrukci intervalů nejprve F. a následně, od nějaké hodnoty skóre F0.. Jako vhodná hodnota skóre pro tento účel se jeví hodnota s 0, ve které se protínají hustoty skóre, rozdíl distribučních funkcí skóre nabývá své maximální hodnoty a také platí, že funkce f IV nabývá nulové hodnoty., s 0 Point of intersection of densities = = Point of maximal difference of CDFs Point of zero value of f IV Finanční matematika v praxi I, Jedovnice 20 22/30
24 ESIS.2, Algorithm for the modified ESIS: ) 2) 3) 4) s0 arg max F q j j k n F, j,, F ( 0 ) n k j k n0 n F0 j F0 ( s0),, n0 k k s q j s F 0 0 0, q [min( score ), q, q0, max( score ) ] 5) Merge intervals given by q where number of bads is less than k. 6) Merge intervals given by q 0 where number of goods is less than k. Finanční matematika v praxi I, Jedovnice 20 ˆ I val, ESIS 2 where 23/30
25 and AESIS.2 Simulation results Consider 000, 0000 and clients, 00p B % of bad clients with f0 : N( 0,) and 00(-p B )% of good clients with f : N(,). Set 0 0, 5 0.5,and. and consider. E(( Iˆ val I val ) 2 ) n k p B n n k pb Finanční matematika v praxi I, Jedovnice 20 0 k p B 24/30
26 Simulation results Dependence of on k. n 000, p 0.2 B n 00000, p 0.05 B n 0000, p 0.2 B pb n k ˆ ˆ0 ˆ I val, AESIS 2 2 n 0000 pb n k ˆ ˆ0 0 2 p B k 0 p B /30
27 Simulation results n=000, n=000, =.0 0 = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, , , , IV_esis 0,0279 0, , , IV_aesis 0, , , ,00903 IV_esis2 0,0382 0, , , IV_esis2a 0, , ,044 0, IV_esis2b 0, , , ,04985 IV_aesis2 0,0570 0, ,043 0, esis2 esis2a eis2b aesis2 algoritmus ESIS.2 ESIS.2 ESIS.2 ESIS.2 Finanční matematika v praxi I, Jedovnice 20 k k pb n 3. ˆ ˆ0 4 pb n k ˆ ˆ0 2 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_aesis 0,2568 0, , , IV_esis 0, ,4470 0,5949 0, IV_aesis 0, , , , IV_esis2 0,7946 0, , ,02286 IV_esis2a 0,2349 0, , ,03476 IV_esis2b 0, , , , IV_aesis2 0, ,0964 0, , n=000, =.5 0 worst average best performance IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 IV_aesis 0, , ,3587 0, IV_esis,50276, , ,86092 IV_aesis 0, , ,2057 0,3756 IV_esis2 0, , , ,6534 IV_esis2a 0, ,8388 0, , IV_esis2b 0, ,260 0, ,09247 IV_aesis2 0, , , , /30
28 Simulation results n=00000, = 0.5 n=00000, =.0 0 IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 IV_aesis 0, , , , IV_esis 0, , , ,00079 IV_aesis 0, , , ,00023 IV_esis2 0, , , ,0003 IV_esis2a 0, , , , IV_esis2b 0,0052 0, , ,00039 IV_aesis2 0, ,0002 0,0009 0, IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, ,0057 0, ,0003 IV_esis 0, , , ,034 IV_aesis 0, , , ,00534 IV_esis2 0, , , , IV_esis2a 0, ,004 0, ,00035 IV_esis2b 0, , , , IV_aesis2 0, , , , worst average best performance Finanční matematika v praxi I, Jedovnice 20 n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , IV_aesis 0, , , ,0472 IV_esis 0, , ,297 0,6950 IV_aesis 0, , , , IV_esis2 0,0258 0, , , IV_esis2a 0, , , , IV_esis2b 0,0927 0, , ,00373 IV_aesis2 0, , , , /30
29 . Simulation results n=000, = r IV_kern 0, , , ,08946 IV_esis 0, , , , IV_aesis2 0, , , ,03350 n=000, =.0 0 r IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_esis2 0,7946 0, , ,02286 n=000, =.5 0 r IV_kern 0, , ,8628 0, IV_esis2a 0, ,0847 0, , IV_aesis2 0, ,0825 0, ,0328 n=00000, = r IV_kern 0, , , , IV_esis 0, ,0009 0,0099 0,00075 IV_aesis2 0, , , , n=00000 =.0 0 r IV_kern 0, , , , IV_esis 0, , , , IV_esis2 0, , , , n=00000 =.5 0 r IV_kern 0, , , ,0026 IV_esis2a 0, , , , IV_aesis2 0, , , , Relativní (r) jde o z předchozích tabulek vydělené příslušnou teoretickou hodnotou IV. Umožní lepší porovnání eliminuje vliv absolutní výše IV. Finanční matematika v praxi I, Jedovnice 20 28/30
30 Conclusions The classical way of computation of the information value, i.e. empirical estimate using deciles of scores, may lead to strongly biased results. We conclude that kernel estimates and empirical estimates with supervised interval selection (ESIS, AESIS, ESIS., ESIS.2 and AESIS.2) are much more appropriate to use. Finanční matematika v praxi I, Jedovnice 20 29/30
31 Děkuji za pozornost. Finanční matematika v praxi I, Jedovnice 20 30/30
INFORMATION VALUE ESTIMATOR FOR CREDIT SCORING MODELS
ECDM Lisbon INFORMATION VALUE ESTIMATOR FOR CREDIT SCORING MODELS Martin Řezáč Dept. of Mathematics and Statistics, Faculty of Science, Masaryk University Introduction Information value is widely used
More informationJádrové odhady regresní funkce pro korelovaná data
Jádrové odhady regresní funkce pro korelovaná data Ústav matematiky a statistiky MÚ Brno Finanční matematika v praxi III., Podlesí 3.9.-4.9. 2013 Obsah Motivace Motivace Motivace Co se snažíme získat?
More informationJádrové odhady gradientu regresní funkce
Monika Kroupová Ivana Horová Jan Koláček Ústav matematiky a statistiky, Masarykova univerzita, Brno ROBUST 2018 Osnova Regresní model a odhad gradientu Metody pro odhad vyhlazovací matice Simulace Závěr
More informationAn Introduction to Relative Distribution Methods
An Introduction to Relative Distribution Methods by Mark S Handcock Professor of Statistics and Sociology Center for Statistics and the Social Sciences CSDE Seminar Series March 2, 2001 In collaboration
More informationMinimum Hellinger Distance Estimation in a. Semiparametric Mixture Model
Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.
More informationSUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION
SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology
More informationMATH 580, exam II SOLUTIONS
MATH 580, exam II SOLUTIONS Solve the problems in the spaces provided on the question sheets. If you run out of room for an answer, continue on the back of the page or use the Extra space empty page at
More informationFinding Divisors, Number of Divisors and Summation of Divisors. Jane Alam Jan
Finding Divisors, Number of Divisors and Summation of Divisors by Jane Alam Jan Finding Divisors If we observe the property of the numbers, we will find some interesting facts. Suppose we have a number
More informationSegmentace textury. Jan Kybic
Segmentace textury Případová studie Jan Kybic Zadání Mikroskopický obrázek segmentujte do tříd: Příčná vlákna Podélná vlákna Matrice Trhliny Zvolená metoda Deskriptorový popis Učení s učitelem ML klasifikátor
More informationLearning Objectives. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 7.2, Page 1
Learning Objectives At the end of the class you should be able to: identify a supervised learning problem characterize how the prediction is a function of the error measure avoid mixing the training and
More informationLecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019
Lecture 10: Probability distributions DANIEL WELLER TUESDAY, FEBRUARY 19, 2019 Agenda What is probability? (again) Describing probabilities (distributions) Understanding probabilities (expectation) Partial
More information12 - Nonparametric Density Estimation
ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6
More informationMeasures of Location
Chapter 7 Measures of Location Definition of Measures of Location (page 219) A measure of location provides information on the percentage of observations in the collection whose values are less than or
More informationMiloš Kopa. Decision problems with stochastic dominance constraints
Decision problems with stochastic dominance constraints Motivation Portfolio selection model Mean risk models max λ Λ m(λ r) νr(λ r) or min λ Λ r(λ r) s.t. m(λ r) µ r is a random vector of assets returns
More informationDescriptive Statistics
Descriptive Statistics DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Descriptive statistics Techniques to visualize
More informationModeling Uncertainty in the Earth Sciences Jef Caers Stanford University
Probability theory and statistical analysis: a review Modeling Uncertainty in the Earth Sciences Jef Caers Stanford University Concepts assumed known Histograms, mean, median, spread, quantiles Probability,
More informationFrequency Estimation of Rare Events by Adaptive Thresholding
Frequency Estimation of Rare Events by Adaptive Thresholding J. R. M. Hosking IBM Research Division 2009 IBM Corporation Motivation IBM Research When managing IT systems, there is a need to identify transactions
More informationA.0 SF s-uncertainty-accuracy-precision
A.0 SF s-uncertainty-accuracy-precision Objectives: Determine the #SF s in a measurement Round a calculated answer to the correct #SF s Round a calculated answer to the correct decimal place Calculate
More informationTrends in the Relative Distribution of Wages by Gender and Cohorts in Brazil ( )
Trends in the Relative Distribution of Wages by Gender and Cohorts in Brazil (1981-2005) Ana Maria Hermeto Camilo de Oliveira Affiliation: CEDEPLAR/UFMG Address: Av. Antônio Carlos, 6627 FACE/UFMG Belo
More informationLoss Estimation using Monte Carlo Simulation
Loss Estimation using Monte Carlo Simulation Tony Bellotti, Department of Mathematics, Imperial College London Credit Scoring and Credit Control Conference XV Edinburgh, 29 August to 1 September 2017 Motivation
More informationRobust Stochastic Frontier Analysis: a Minimum Density Power Divergence Approach
Robust Stochastic Frontier Analysis: a Minimum Density Power Divergence Approach Federico Belotti Giuseppe Ilardi CEIS, University of Rome Tor Vergata Bank of Italy Workshop on the Econometrics and Statistics
More informationFrom Mating Pool Distributions to Model Overfitting
From Mating Pool Distributions to Model Overfitting Claudio F. Lima 1 Fernando G. Lobo 1 Martin Pelikan 2 1 Department of Electronics and Computer Science Engineering University of Algarve, Portugal 2
More informationCALCULATING MAGNETIC FIELDS & THE BIOT-SAVART LAW. Purdue University Physics 241 Lecture 15 Brendan Sullivan
CALCULATING MAGNETIC FIELDS & THE BIOT-SAVAT LAW Purdue University Physics 41 Lecture 15 Brendan Sullivan Introduction Brendan Sullivan, PHYS89, sullivb@purdue.edu Office Hours: By Appointment Just stop
More informationCSCE 750 Final Exam Answer Key Wednesday December 7, 2005
CSCE 750 Final Exam Answer Key Wednesday December 7, 2005 Do all problems. Put your answers on blank paper or in a test booklet. There are 00 points total in the exam. You have 80 minutes. Please note
More informationOverview (Fall 2007) Machine Translation Part III. Roadmap for the Next Few Lectures. Phrase-Based Models. Learning phrases from alignments
Overview Learning phrases from alignments 6.864 (Fall 2007) Machine Translation Part III A phrase-based model Decoding in phrase-based models (Thanks to Philipp Koehn for giving me slides from his EACL
More informationOikos. Appendix 1 and 2. o20751
Oikos o20751 Rosindell, J. and Cornell, S. J. 2013. Universal scaling of species-abundance distributions across multiple scales. Oikos 122: 1101 1111. Appendix 1 and 2 Universal scaling of species-abundance
More informationCOMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017
COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University SEQUENTIAL DATA So far, when thinking
More informationSUMMARIZING MEASURED DATA. Gaia Maselli
SUMMARIZING MEASURED DATA Gaia Maselli maselli@di.uniroma1.it Computer Network Performance 2 Overview Basic concepts Summarizing measured data Summarizing data by a single number Summarizing variability
More informationPROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS
PROBABILITY AND INFORMATION THEORY Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Probability space Rules of probability
More informationProofs and derivations
A Proofs and derivations Proposition 1. In the sheltering decision problem of section 1.1, if m( b P M + s) = u(w b P M + s), where u( ) is weakly concave and twice continuously differentiable, then f
More informationRegression I: Mean Squared Error and Measuring Quality of Fit
Regression I: Mean Squared Error and Measuring Quality of Fit -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD 1 The Setup Suppose there is a scientific problem we are interested in solving
More informationSTATISTICS 1 REVISION NOTES
STATISTICS 1 REVISION NOTES Statistical Model Representing and summarising Sample Data Key words: Quantitative Data This is data in NUMERICAL FORM such as shoe size, height etc. Qualitative Data This is
More informationStar-Structured High-Order Heterogeneous Data Co-clustering based on Consistent Information Theory
Star-Structured High-Order Heterogeneous Data Co-clustering based on Consistent Information Theory Bin Gao Tie-an Liu Wei-ing Ma Microsoft Research Asia 4F Sigma Center No. 49 hichun Road Beijing 00080
More informationDensity Modeling and Clustering Using Dirichlet Diffusion Trees
p. 1/3 Density Modeling and Clustering Using Dirichlet Diffusion Trees Radford M. Neal Bayesian Statistics 7, 2003, pp. 619-629. Presenter: Ivo D. Shterev p. 2/3 Outline Motivation. Data points generation.
More informationSecond main application of Chernoff: analysis of load balancing. Already saw balls in bins example. oblivious algorithms only consider self packet.
Routing Second main application of Chernoff: analysis of load balancing. Already saw balls in bins example synchronous message passing bidirectional links, one message per step queues on links permutation
More informationClass XII_All India_Mathematics_Set-2 SECTION C. Question numbers 23 to 29 carry 6 marks each.
SECTION C Question numbers 3 to 9 carr 6 marks each. 3. Find the equation of the plane passing through the line of intersection of the planes r ˆi 3j ˆ 6 0 r 3i ˆ ˆ j k ˆ 0, whose perpendicular distance
More informationMath 20 Spring Discrete Probability. Midterm Exam
Math 20 Spring 203 Discrete Probability Midterm Exam Thursday April 25, 5:00 7:00 PM Your name (please print): Instructions: This is a closed book, closed notes exam. Use of calculators is not permitted.
More informationLecture 3 Probability Basics
Lecture 3 Probability Basics Thais Paiva STA 111 - Summer 2013 Term II July 3, 2013 Lecture Plan 1 Definitions of probability 2 Rules of probability 3 Conditional probability What is Probability? Probability
More informationAnalysis of Algorithms. Unit 5 - Intractable Problems
Analysis of Algorithms Unit 5 - Intractable Problems 1 Intractable Problems Tractable Problems vs. Intractable Problems Polynomial Problems NP Problems NP Complete and NP Hard Problems 2 In this unit we
More informationEfficient Semiparametric Estimators via Modified Profile Likelihood in Frailty & Accelerated-Failure Models
NIH Talk, September 03 Efficient Semiparametric Estimators via Modified Profile Likelihood in Frailty & Accelerated-Failure Models Eric Slud, Math Dept, Univ of Maryland Ongoing joint project with Ilia
More informationMachine Learning. Lecture 9: Learning Theory. Feng Li.
Machine Learning Lecture 9: Learning Theory Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Why Learning Theory How can we tell
More informationWESTMORELAND COUNTY PUBLIC SCHOOLS Integrated Instructional Pacing Guide and Checklist Algebra II
WESTMORELAND COUNTY PUBLIC SCHOOLS 201 2014 Integrated Instructional Pacing Guide and Checklist Algebra II (s) ESS Vocabulary AII.4 Absolute Value Equations and Inequalities Absolute Value Equations and
More informationComparison of Methods for Computation of Ideal Circulation Distribution on the Propeller
Comparison of Methods for Computation of Ideal Circulation Distribution on the Propeller Ing. Jan Klesa vedoucí práce: doc. Ing. Daniel Hanus, CSc. Abstrakt Ideální rozložení cirkulace na vrtuli je vypočteno
More informationQuantifying Stochastic Model Errors via Robust Optimization
Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations
More informationCurve Fitting Re-visited, Bishop1.2.5
Curve Fitting Re-visited, Bishop1.2.5 Maximum Likelihood Bishop 1.2.5 Model Likelihood differentiation p(t x, w, β) = Maximum Likelihood N N ( t n y(x n, w), β 1). (1.61) n=1 As we did in the case of the
More informationČasopis pro pěstování matematiky a fysiky
Časopis pro pěstování matematiky a fysiky Norman Levinson Criteria for the limit-point case for second order linear differential operators Časopis pro pěstování matematiky a fysiky, Vol. 74 (1949), No.
More informationEE Power Gain and Amplifier Design 10/31/2017
EE 40458 Power Gain and Amplifier Design 10/31/017 Motivation Brief recap: We ve studied matching networks (several types, how to design them, bandwidth, how they work, etc ) Studied network analysis techniques
More informationECS171: Machine Learning
ECS171: Machine Learning Lecture 6: Training versus Testing (LFD 2.1) Cho-Jui Hsieh UC Davis Jan 29, 2018 Preamble to the theory Training versus testing Out-of-sample error (generalization error): What
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationProbability Models for Bayesian Recognition
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIAG / osig Second Semester 06/07 Lesson 9 0 arch 07 Probability odels for Bayesian Recognition Notation... Supervised Learning for Bayesian
More informationRevealed Preference 2011
Revealed Preference 2011 Motivation: 1. up until now we have started with preference and then described behaviour 2. revealed preference works backwards - start with behaviour and describe preferences
More information2.1.3 The Testing Problem and Neave s Step Method
we can guarantee (1) that the (unknown) true parameter vector θ t Θ is an interior point of Θ, and (2) that ρ θt (R) > 0 for any R 2 Q. These are two of Birch s regularity conditions that were critical
More informationEcon 582 Nonparametric Regression
Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume
More informationFINAL: CS 6375 (Machine Learning) Fall 2014
FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for
More informationMath Review Sheet, Fall 2008
1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the
More informationCS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization
CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization Tim Roughgarden February 28, 2017 1 Preamble This lecture fulfills a promise made back in Lecture #1,
More informationDover- Sherborn High School Mathematics Curriculum Probability and Statistics
Mathematics Curriculum A. DESCRIPTION This is a full year courses designed to introduce students to the basic elements of statistics and probability. Emphasis is placed on understanding terminology and
More informationChapter 5-2: Clustering
Chapter 5-2: Clustering Jilles Vreeken Revision 1, November 20 th typo s fixed: dendrogram Revision 2, December 10 th clarified: we do consider a point x as a member of its own ε-neighborhood 12 Nov 2015
More information1 Fields and vector spaces
1 Fields and vector spaces In this section we revise some algebraic preliminaries and establish notation. 1.1 Division rings and fields A division ring, or skew field, is a structure F with two binary
More informationSO x is a cubed root of t
7.6nth Roots 1) What do we know about x because of the following equation x 3 = t? All in one.docx SO x is a cubed root of t 2) Definition of nth root: 3) Study example 1 4) Now try the following problem
More informationLecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron
CS446: Machine Learning, Fall 2017 Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron Lecturer: Sanmi Koyejo Scribe: Ke Wang, Oct. 24th, 2017 Agenda Recap: SVM and Hinge loss, Representer
More informationCoding on Countably Infinite Alphabets
Coding on Countably Infinite Alphabets Non-parametric Information Theory Licence de droits d usage Outline Lossless Coding on infinite alphabets Source Coding Universal Coding Infinite Alphabets Enveloppe
More informationDescribing Distributions with Numbers
Topic 2 We next look at quantitative data. Recall that in this case, these data can be subject to the operations of arithmetic. In particular, we can add or subtract observation values, we can sort them
More informationIf something is repeated, look for what stays unchanged!
LESSON 5: INVARIANTS Invariants are very useful in problems concerning algorithms involving steps that are repeated again and again (games, transformations). While these steps are repeated, the system
More informationORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES
ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES 1. Order statistics Let X 1,...,X n be n real-valued observations. One can always arrangetheminordertogettheorder statisticsx (1) X (2) X (n). SinceX (k)
More informationKullback-Leibler Designs
Kullback-Leibler Designs Astrid JOURDAN Jessica FRANCO Contents Contents Introduction Kullback-Leibler divergence Estimation by a Monte-Carlo method Design comparison Conclusion 2 Introduction Computer
More informationSo far we discussed random number generators that need to have the maximum length period.
So far we discussed random number generators that need to have the maximum length period. Even the increment series has the maximum length period yet it is by no means random How do we decide if a generator
More informationIntermediate Math Circles February 14, 2018 Contest Prep: Number Theory
Intermediate Math Circles February 14, 2018 Contest Prep: Number Theory Part 1: Prime Factorization A prime number is an integer greater than 1 whose only positive divisors are 1 and itself. An integer
More informationMAT 146. Semester Exam Part II 100 points (Part II: 50 points) Calculator Used Impact on Course Grade: approximately 30% Score
MAT 146 Semester Exam Part II Name 100 points (Part II: 50 points) Calculator Used Impact on Course Grade: approximately 30% Score Questions (17) through (26) are each worth 5 points. See the grading rubric
More informationAlgebra II Unit Breakdown (Curriculum Map Outline)
QUARTER 1 Unit 1 (Arithmetic and Geometric Sequences) A. Sequences as Functions 1. Identify finite and infinite sequences 2. Identify an arithmetic sequence and its parts 3. Identify a geometric sequence
More informationJohns Hopkins University, Department of Mathematics Abstract Algebra - Spring 2013 Midterm Exam Solution
Johns Hopkins University, Department of Mathematics 110.40 Abstract Algebra - Spring 013 Midterm Exam Solution Instructions: This exam has 6 pages. No calculators, books or notes allowed. You must answer
More informationDept of Computer Science, Michigan State University b
CONTOUR REGRESSION: A distribution-regularized regression framework for climate modeling Zubin Abraham a, Pang-Ning Tan a, Julie A. Winkler b, Perdinan b, Shiyuan Zhong b, Malgorzata Liszewska c a Dept
More informationIndian Statistical Institute
Indian Statistical Institute Introductory Computer programming Robust Regression methods with high breakdown point Author: Roll No: MD1701 February 24, 2018 Contents 1 Introduction 2 2 Criteria for evaluating
More informationPredicting the Probability of Correct Classification
Predicting the Probability of Correct Classification Gregory Z. Grudic Department of Computer Science University of Colorado, Boulder grudic@cs.colorado.edu Abstract We propose a formulation for binary
More informationCS570 Introduction to Data Mining
CS570 Introduction to Data Mining Department of Mathematics and Computer Science Li Xiong Data Exploration and Data Preprocessing Data and Attributes Data exploration Data pre-processing Data cleaning
More informationVers un apprentissage subquadratique pour les mélanges d arbres
Vers un apprentissage subquadratique pour les mélanges d arbres F. Schnitzler 1 P. Leray 2 L. Wehenkel 1 fschnitzler@ulg.ac.be 1 Université deliège 2 Université de Nantes 10 mai 2010 F. Schnitzler (ULG)
More informationFall 2014 CMSC250/250H Midterm II
Fall 2014 CMSC250/250H Midterm II Circle Your Section! 0101 (10am: 3120, Ladan) 0102 (11am: 3120, Ladan) 0103 (Noon: 3120, Peter) 0201 (2pm: 3120, Yi) 0202 (10am: 1121, Vikas) 0203 (11am: 1121, Vikas)
More informationPattern Popularity in 132-Avoiding Permutations
Pattern Popularity in 132-Avoiding Permutations The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher Rudolph,
More informationBias in RNA sequencing and what to do about it
Bias in RNA sequencing and what to do about it Walter L. (Larry) Ruzzo Computer Science and Engineering Genome Sciences University of Washington Fred Hutchinson Cancer Research Center Seattle, WA, USA
More informationLatent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology
Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2016 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,
More informationENSEMBLE FLOOD INUNDATION FORECASTING: A CASE STUDY IN THE TIDAL DELAWARE RIVER
ENSEMBLE FLOOD INUNDATION FORECASTING: A CASE STUDY IN THE TIDAL DELAWARE RIVER Michael Gomez & Alfonso Mejia Civil and Environmental Engineering Pennsylvania State University 10/12/2017 Mid-Atlantic Water
More information(1) Sort all time observations from least to greatest, so that the j th and (j + 1) st observations are ordered by t j t j+1 for all j = 1,..., J.
AFFIRMATIVE ACTION AND HUMAN CAPITAL INVESTMENT 8. ONLINE APPENDIX TO ACCOMPANY Affirmative Action and Human Capital Investment: Theory and Evidence from a Randomized Field Experiment, by CHRISTOPHER COTTON,
More informationQUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS
QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS Parvathinathan Venkitasubramaniam, Gökhan Mergen, Lang Tong and Ananthram Swami ABSTRACT We study the problem of quantization for
More informationData Mining Techniques
Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!
More informationFEEG6017 lecture: Akaike's information criterion; model reduction. Brendan Neville
FEEG6017 lecture: Akaike's information criterion; model reduction Brendan Neville bjn1c13@ecs.soton.ac.uk Occam's razor William of Occam, 1288-1348. All else being equal, the simplest explanation is the
More informationUnit Two Descriptive Biostatistics. Dr Mahmoud Alhussami
Unit Two Descriptive Biostatistics Dr Mahmoud Alhussami Descriptive Biostatistics The best way to work with data is to summarize and organize them. Numbers that have not been summarized and organized are
More informationMATHEMATICS (New Specification)
MS WELSH JOINT EDUCATION COMMITTEE. CYD-BWYLLGOR ADDYSG CYMRU General Certificate of Education Advanced Subsidiary/Advanced Tystysgrif Addysg Gyffredinol Uwch Gyfrannol/Uwch MARKING SCHEMES JANUARY 6 MATHEMATICS
More informationCOMS 4705, Fall Machine Translation Part III
COMS 4705, Fall 2011 Machine Translation Part III 1 Roadmap for the Next Few Lectures Lecture 1 (last time): IBM Models 1 and 2 Lecture 2 (today): phrase-based models Lecture 3: Syntax in statistical machine
More informationUnit 6 Study Guide: Equations. Section 6-1: One-Step Equations with Adding & Subtracting
Unit 6 Study Guide: Equations DUE DATE: A Day: Dec 18 th B Day: Dec 19 th Name Period Score / Section 6-1: One-Step Equations with Adding & Subtracting Textbook Reference: Page 437 Vocabulary: Equation
More informationA short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie
A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie Computational Biology Program Memorial Sloan-Kettering Cancer Center http://cbio.mskcc.org/leslielab
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationBonding in solids The interaction of electrons in neighboring atoms of a solid serves the very important function of holding the crystal together.
Bonding in solids The interaction of electrons in neighboring atoms of a solid serves the very important function of holding the crystal together. For example Nacl In the Nacl lattice, each Na atom is
More informationLecture 3: More on regularization. Bayesian vs maximum likelihood learning
Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting
More informationDecision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag
Decision Trees Nicholas Ruozzi University of Texas at Dallas Based on the slides of Vibhav Gogate and David Sontag Supervised Learning Input: labelled training data i.e., data plus desired output Assumption:
More informationPollution Sources Detection via Principal Component Analysis and Rotation
Pollution Sources Detection via Principal Component Analysis and Rotation Vanessa Kuentz 1 in collaboration with : Marie Chavent 1 Hervé Guégan 2 Brigitte Patouille 1 Jérôme Saracco 1,3 1 IMB, Université
More informationMATH 120. Test 1 Spring, 2012 DO ALL ASSIGNED PROBLEMS. Things to particularly study
MATH 120 Test 1 Spring, 2012 DO ALL ASSIGNED PROBLEMS Things to particularly study 1) Critical Thinking Basic strategies Be able to solve using the basic strategies, such as finding patterns, questioning,
More informationLecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016
Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,
More informationEstimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk
Ann Inst Stat Math (0) 64:359 37 DOI 0.007/s0463-00-036-3 Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Paul Vos Qiang Wu Received: 3 June 009 / Revised:
More informationIntermediate Algebra with Applications
Lakeshore Technical College 10-804-118 Intermediate Algebra with Applications Course Outcome Summary Course Information Alternate Title Description Total Credits 4 Total Hours 72 Pre/Corequisites Prerequisite
More informationAlgebra 2 College Prep Curriculum Maps
Algebra 2 College Prep Curriculum Maps Unit 1: Polynomial, Rational, and Radical Relationships Unit 2: Modeling With Functions Unit 3: Inferences and Conclusions from Data Unit 4: Trigonometric Functions
More information