Computation of Information Value for Credit Scoring Models

Size: px
Start display at page:

Download "Computation of Information Value for Credit Scoring Models"

Transcription

1 Jedovnice 20 Computation of Information Value for Credit Scoring Models Martin Řezáč, Jan Koláček Dept. of Mathematics and Statistics, Faculty of Science, Masaryk University

2 Information value The special case of Kullback-Leibler divergence given by: where are densities of scores of bad and good clients. Finanční matematika v praxi I, Jedovnice 20 /30

3 Kernel estimate The kernel density estimates are defined by where and K is the Epanechnikov kernel. Finanční matematika v praxi I, Jedovnice 20 2/30

4 Empirical estimate of I val Finanční matematika v praxi I, Jedovnice 20 3/30

5 Empirical estimate of I val However in practice, there could occur computational problems. The Information value index becomes infinite in cases when some of n 0 j or n j are equal to 0. Choosing of the number of bins is also very important. In the literature and also in many applications in credit scoring, the value r=0 is preferred. Finanční matematika v praxi I, Jedovnice 20 4/30

6 Empirical estimate with supervised interval selection (ESIS), We want to avoid zero values of n 0 j or n j. We propose to require to have at least k, where k is a positive integer, observations of scores of both good and bad client in each interval. This is the basic idea of all proposed algorithms. Finanční matematika v praxi I, Jedovnice 20 5/30

7 Empirical estimate with supervised interval selection, the first ESIS: Set where is the empirical quantile function appropriate to the empirical cumulative distribution function of scores of bad clients. Finanční matematika v praxi I, Jedovnice 20 6/30

8 Empirical estimate with supervised interval selection Usage of quantile function of scores of bad clients is motivated by the assumption, that number of bad clients is less than number of good clients. If n 0 is not divisible by k, it is necessary to adjust our intervals, because we obtain number of scores of bad clients in the last interval, which is less than k. In this case, we have to merge the last two intervals. Furthermore we need to ensure, that the number of scores of good clients is as required in each interval. To do so, we compute n j for all actual intervals. If we obtain n j < k for jth interval, we merge this interval with its neighbor on the right side. This can be done for all intervals except the last one. If we have n < k for the j last interval, than we have to merge it with its neighbor on the left side, i.e. we merge the last two intervals. Finanční matematika v praxi I, Jedovnice 20 7/30

9 and Empirical estimate with supervised interval selection Very important is the choice of k. If we choose too small value, we get overestimated value of the Information value, and vice versa. As a reasonable compromise seems to be adjusted square root of number of bad clients given by The estimate of the Information value is given by where n and n 0 correspond to observed counts of good and bad clients j j in intervals created according to the described procedure. Finanční matematika v praxi I, Jedovnice 20 8/30

10 Simulation results Consider n clients, 00p B % of bad clients with and 00(-p B )% of good clients with f N(, ). : 0 Because of normality we know I val. Consider following values of parameters: n = , n = 000 μ 0 = 0 σ 0 = σ = μ = 0.5,,.5 p B = 0.02, 0.05, 0., f : N( 0, 0) 0 Finanční matematika v praxi I, Jedovnice 20 9/30

11 Simulation results ) Scores of bad and good clients were generated according to given parameters. 2) Estimates ˆ, Iˆ, Iˆ were computed. 3) Square errors were computed. 4) Steps )-3) were repeated one thousand times. 5) was computed. I val, DEC val, KERN val, ESIS Finanční matematika v praxi I, Jedovnice 20 0/30

12 Simulation results n=00000, = IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 n=00000, =.0 0 IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , n=000, = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , n=000, =.0 0 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , n=000, =.5 0 IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 worst average best performance Finanční matematika v praxi I, Jedovnice 20 /30

13 and Adjusted empirical estimate with supervised interval selection (AESIS) It is obvious that the choice of parameter k is crucial. So the question is: Is the choice optimal (according to )? What effect has n 0 on the optimal k? And what effect, if any, has the difference of means? 0 Finanční matematika v praxi I, Jedovnice 20 2/30

14 Simulation results Consider 0000 clients, 00p B % of bad clients with 00(-p B )% of good clients with f : N(,). Set 0 0 and consider 0.5,and.5,. E(( Iˆ val I val ) 2 ) f : N( 0,) 0 and k p B p B 0 0 3/30

15 Simulation results Dependence of on k,. 0 The highlighted circles correspond to values of k, where minimal value of the is obtained. The diamonds correspond to values of k given by. 0 p B k 0 p B ˆ I val, AESIS 4/30

16 Simulation results n=000, = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, , , , n=000, =.0 0 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_aesis 0,2568 0, , , n=000, =.5 0 The classical estimate had the lowest in case of weak model and extremely low number (namely 20) of bad clients. If the number of bad client is slightly higher, had the best performance. Considering models with higher predictive power, and had the best performance. IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 IV_aesis 0, , ,3587 0, Finanční matematika v praxi I, Jedovnice 20 5/30

17 Simulation results n=00000, = IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 IV_aesis 0, , , , n=00000, =.0 0 IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, ,0057 0, ,0003 n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , IV_aesis 0, , , ,0472 We can see that the classical estimate was outperformed by all other estimates of the Information value in all considered cases when the number of clients was high (00000 in our case). The best performance was achieved by in case of weak scoring models, by in case of models with high performance and by for models with very high performance. Finanční matematika v praxi I, Jedovnice 20 5/30

18 Algorithm for the modified ESIS: ) 2) 3) q [] k q j F n smax max( q j, q j 0 ) q j 0 F 0 k n ESIS. 0, ˆ I val, ESIS where s max 4) Add to the sequence, i.e. q [ q, smax ] 5) Erase all scores s max 6) While n 0 and n are greater than 2*k, repeat step 2) 5) 7) q [min( score ), q] Finanční matematika v praxi I, Jedovnice 20 7/30

19 and AESIS. Simulation results Consider 000 and 0000 clients, 00p B % of bad clients with and 00(-p B )% of good clients with f : N(,). Set 0 0 and consider 0.5,and.5,. E(( Iˆ val I val ) 2 ) f : N( 0,) 0 n 000 n 0000 k p B k pb Finanční matematika v praxi I, Jedovnice 20 8/30

20 Simulation results Dependence of on k. n 000, p 0.2 B n 0000, p 0.2 B k pb n 3. 4 ˆ ˆ0 n 000 pb n k 3. 4 ˆ ˆ0 0 n 0000 p B k 0 p B ˆ I val, AESIS pb n k 3. 4 ˆ ˆ0 0 p B k 0 p B /30

21 Simulation results n=00000, = IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 IV_aesis 0, , , , IV_esis 0, , , ,00079 IV_aesis 0, , , ,00023 n=00000, =.0 0 IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, ,0057 0, ,0003 IV_esis 0, , , ,034 IV_aesis 0, , , ,00534 n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , IV_aesis 0, , , ,0472 IV_esis 0, , ,297 0,6950 IV_aesis 0, , , , Finanční matematika v praxi I, Jedovnice 20 n=000, = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, , , , IV_esis 0,0279 0, , , IV_aesis 0, , , ,00903 n=000, =.0 0 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_aesis 0,2568 0, , , IV_esis 0, ,4470 0,5949 0, IV_aesis 0, , , , n=000, =.5 0 IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 IV_aesis 0, , ,3587 0, IV_esis,50276, , ,86092 IV_aesis 0, , ,2057 0, /30

22 . Simulation results The new algorithm (ESIS.) ended disastrously in most cases. The only exception was a situation where n=000 a μ -μ 0 =0.5. This corresponds to a scoring model with very poor discriminatory power and a very small number of observed bad clients (20-200). AESIS. ranked among the average. Finanční matematika v praxi I, Jedovnice 20 2/30

23 ESIS.2 U původního ESIS často dochází ke slučování vypočtených intervalů ve druhé fázi algoritmu. Pro výpočet se používá jen F0.. Aby byla splněna podmínka n >k, je zřejmě nutné, aby hranice prvního intervalu byla větší než k. F n To vede k myšlence použít ke konstrukci intervalů nejprve F. a následně, od nějaké hodnoty skóre F0.. Jako vhodná hodnota skóre pro tento účel se jeví hodnota s 0, ve které se protínají hustoty skóre, rozdíl distribučních funkcí skóre nabývá své maximální hodnoty a také platí, že funkce f IV nabývá nulové hodnoty., s 0 Point of intersection of densities = = Point of maximal difference of CDFs Point of zero value of f IV Finanční matematika v praxi I, Jedovnice 20 22/30

24 ESIS.2, Algorithm for the modified ESIS: ) 2) 3) 4) s0 arg max F q j j k n F, j,, F ( 0 ) n k j k n0 n F0 j F0 ( s0),, n0 k k s q j s F 0 0 0, q [min( score ), q, q0, max( score ) ] 5) Merge intervals given by q where number of bads is less than k. 6) Merge intervals given by q 0 where number of goods is less than k. Finanční matematika v praxi I, Jedovnice 20 ˆ I val, ESIS 2 where 23/30

25 and AESIS.2 Simulation results Consider 000, 0000 and clients, 00p B % of bad clients with f0 : N( 0,) and 00(-p B )% of good clients with f : N(,). Set 0 0, 5 0.5,and. and consider. E(( Iˆ val I val ) 2 ) n k p B n n k pb Finanční matematika v praxi I, Jedovnice 20 0 k p B 24/30

26 Simulation results Dependence of on k. n 000, p 0.2 B n 00000, p 0.05 B n 0000, p 0.2 B pb n k ˆ ˆ0 ˆ I val, AESIS 2 2 n 0000 pb n k ˆ ˆ0 0 2 p B k 0 p B /30

27 Simulation results n=000, n=000, =.0 0 = IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, , , , IV_esis 0,0279 0, , , IV_aesis 0, , , ,00903 IV_esis2 0,0382 0, , , IV_esis2a 0, , ,044 0, IV_esis2b 0, , , ,04985 IV_aesis2 0,0570 0, ,043 0, esis2 esis2a eis2b aesis2 algoritmus ESIS.2 ESIS.2 ESIS.2 ESIS.2 Finanční matematika v praxi I, Jedovnice 20 k k pb n 3. ˆ ˆ0 4 pb n k ˆ ˆ0 2 IV_decil 0, , , , IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_aesis 0,2568 0, , , IV_esis 0, ,4470 0,5949 0, IV_aesis 0, , , , IV_esis2 0,7946 0, , ,02286 IV_esis2a 0,2349 0, , ,03476 IV_esis2b 0, , , , IV_aesis2 0, ,0964 0, , n=000, =.5 0 worst average best performance IV_decil,663859, , , IV_kern 0, , , ,96856 IV_esis 0, ,3525 0,7293 0,94676 IV_aesis 0, , ,3587 0, IV_esis,50276, , ,86092 IV_aesis 0, , ,2057 0,3756 IV_esis2 0, , , ,6534 IV_esis2a 0, ,8388 0, , IV_esis2b 0, ,260 0, ,09247 IV_aesis2 0, , , , /30

28 Simulation results n=00000, = 0.5 n=00000, =.0 0 IV_decil 0, , , ,00068 IV_kern 0, , ,0003 0, IV_esis 0, , , ,00027 IV_aesis 0, , , , IV_esis 0, , , ,00079 IV_aesis 0, , , ,00023 IV_esis2 0, , , ,0003 IV_esis2a 0, , , , IV_esis2b 0,0052 0, , ,00039 IV_aesis2 0, ,0002 0,0009 0, IV_decil 0, , , , IV_kern 0, , , , IV_esis 0, , , , IV_aesis 0, ,0057 0, ,0003 IV_esis 0, , , ,034 IV_aesis 0, , , ,00534 IV_esis2 0, , , , IV_esis2a 0, ,004 0, ,00035 IV_esis2b 0, , , , IV_aesis2 0, , , , worst average best performance Finanční matematika v praxi I, Jedovnice 20 n=00000, =.5 0 IV_decil 0, , , ,02066 IV_kern 0,0956 0, , , IV_esis 0, , , , IV_aesis 0, , , ,0472 IV_esis 0, , ,297 0,6950 IV_aesis 0, , , , IV_esis2 0,0258 0, , , IV_esis2a 0, , , , IV_esis2b 0,0927 0, , ,00373 IV_aesis2 0, , , , /30

29 . Simulation results n=000, = r IV_kern 0, , , ,08946 IV_esis 0, , , , IV_aesis2 0, , , ,03350 n=000, =.0 0 r IV_kern 0,7382 0, , ,0323 IV_esis 0,5088 0, , , IV_esis2 0,7946 0, , ,02286 n=000, =.5 0 r IV_kern 0, , ,8628 0, IV_esis2a 0, ,0847 0, , IV_aesis2 0, ,0825 0, ,0328 n=00000, = r IV_kern 0, , , , IV_esis 0, ,0009 0,0099 0,00075 IV_aesis2 0, , , , n=00000 =.0 0 r IV_kern 0, , , , IV_esis 0, , , , IV_esis2 0, , , , n=00000 =.5 0 r IV_kern 0, , , ,0026 IV_esis2a 0, , , , IV_aesis2 0, , , , Relativní (r) jde o z předchozích tabulek vydělené příslušnou teoretickou hodnotou IV. Umožní lepší porovnání eliminuje vliv absolutní výše IV. Finanční matematika v praxi I, Jedovnice 20 28/30

30 Conclusions The classical way of computation of the information value, i.e. empirical estimate using deciles of scores, may lead to strongly biased results. We conclude that kernel estimates and empirical estimates with supervised interval selection (ESIS, AESIS, ESIS., ESIS.2 and AESIS.2) are much more appropriate to use. Finanční matematika v praxi I, Jedovnice 20 29/30

31 Děkuji za pozornost. Finanční matematika v praxi I, Jedovnice 20 30/30

INFORMATION VALUE ESTIMATOR FOR CREDIT SCORING MODELS

INFORMATION VALUE ESTIMATOR FOR CREDIT SCORING MODELS ECDM Lisbon INFORMATION VALUE ESTIMATOR FOR CREDIT SCORING MODELS Martin Řezáč Dept. of Mathematics and Statistics, Faculty of Science, Masaryk University Introduction Information value is widely used

More information

Jádrové odhady regresní funkce pro korelovaná data

Jádrové odhady regresní funkce pro korelovaná data Jádrové odhady regresní funkce pro korelovaná data Ústav matematiky a statistiky MÚ Brno Finanční matematika v praxi III., Podlesí 3.9.-4.9. 2013 Obsah Motivace Motivace Motivace Co se snažíme získat?

More information

Jádrové odhady gradientu regresní funkce

Jádrové odhady gradientu regresní funkce Monika Kroupová Ivana Horová Jan Koláček Ústav matematiky a statistiky, Masarykova univerzita, Brno ROBUST 2018 Osnova Regresní model a odhad gradientu Metody pro odhad vyhlazovací matice Simulace Závěr

More information

An Introduction to Relative Distribution Methods

An Introduction to Relative Distribution Methods An Introduction to Relative Distribution Methods by Mark S Handcock Professor of Statistics and Sociology Center for Statistics and the Social Sciences CSDE Seminar Series March 2, 2001 In collaboration

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology

More information

MATH 580, exam II SOLUTIONS

MATH 580, exam II SOLUTIONS MATH 580, exam II SOLUTIONS Solve the problems in the spaces provided on the question sheets. If you run out of room for an answer, continue on the back of the page or use the Extra space empty page at

More information

Finding Divisors, Number of Divisors and Summation of Divisors. Jane Alam Jan

Finding Divisors, Number of Divisors and Summation of Divisors. Jane Alam Jan Finding Divisors, Number of Divisors and Summation of Divisors by Jane Alam Jan Finding Divisors If we observe the property of the numbers, we will find some interesting facts. Suppose we have a number

More information

Segmentace textury. Jan Kybic

Segmentace textury. Jan Kybic Segmentace textury Případová studie Jan Kybic Zadání Mikroskopický obrázek segmentujte do tříd: Příčná vlákna Podélná vlákna Matrice Trhliny Zvolená metoda Deskriptorový popis Učení s učitelem ML klasifikátor

More information

Learning Objectives. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 7.2, Page 1

Learning Objectives. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 7.2, Page 1 Learning Objectives At the end of the class you should be able to: identify a supervised learning problem characterize how the prediction is a function of the error measure avoid mixing the training and

More information

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019 Lecture 10: Probability distributions DANIEL WELLER TUESDAY, FEBRUARY 19, 2019 Agenda What is probability? (again) Describing probabilities (distributions) Understanding probabilities (expectation) Partial

More information

12 - Nonparametric Density Estimation

12 - Nonparametric Density Estimation ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6

More information

Measures of Location

Measures of Location Chapter 7 Measures of Location Definition of Measures of Location (page 219) A measure of location provides information on the percentage of observations in the collection whose values are less than or

More information

Miloš Kopa. Decision problems with stochastic dominance constraints

Miloš Kopa. Decision problems with stochastic dominance constraints Decision problems with stochastic dominance constraints Motivation Portfolio selection model Mean risk models max λ Λ m(λ r) νr(λ r) or min λ Λ r(λ r) s.t. m(λ r) µ r is a random vector of assets returns

More information

Descriptive Statistics

Descriptive Statistics Descriptive Statistics DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Descriptive statistics Techniques to visualize

More information

Modeling Uncertainty in the Earth Sciences Jef Caers Stanford University

Modeling Uncertainty in the Earth Sciences Jef Caers Stanford University Probability theory and statistical analysis: a review Modeling Uncertainty in the Earth Sciences Jef Caers Stanford University Concepts assumed known Histograms, mean, median, spread, quantiles Probability,

More information

Frequency Estimation of Rare Events by Adaptive Thresholding

Frequency Estimation of Rare Events by Adaptive Thresholding Frequency Estimation of Rare Events by Adaptive Thresholding J. R. M. Hosking IBM Research Division 2009 IBM Corporation Motivation IBM Research When managing IT systems, there is a need to identify transactions

More information

A.0 SF s-uncertainty-accuracy-precision

A.0 SF s-uncertainty-accuracy-precision A.0 SF s-uncertainty-accuracy-precision Objectives: Determine the #SF s in a measurement Round a calculated answer to the correct #SF s Round a calculated answer to the correct decimal place Calculate

More information

Trends in the Relative Distribution of Wages by Gender and Cohorts in Brazil ( )

Trends in the Relative Distribution of Wages by Gender and Cohorts in Brazil ( ) Trends in the Relative Distribution of Wages by Gender and Cohorts in Brazil (1981-2005) Ana Maria Hermeto Camilo de Oliveira Affiliation: CEDEPLAR/UFMG Address: Av. Antônio Carlos, 6627 FACE/UFMG Belo

More information

Loss Estimation using Monte Carlo Simulation

Loss Estimation using Monte Carlo Simulation Loss Estimation using Monte Carlo Simulation Tony Bellotti, Department of Mathematics, Imperial College London Credit Scoring and Credit Control Conference XV Edinburgh, 29 August to 1 September 2017 Motivation

More information

Robust Stochastic Frontier Analysis: a Minimum Density Power Divergence Approach

Robust Stochastic Frontier Analysis: a Minimum Density Power Divergence Approach Robust Stochastic Frontier Analysis: a Minimum Density Power Divergence Approach Federico Belotti Giuseppe Ilardi CEIS, University of Rome Tor Vergata Bank of Italy Workshop on the Econometrics and Statistics

More information

From Mating Pool Distributions to Model Overfitting

From Mating Pool Distributions to Model Overfitting From Mating Pool Distributions to Model Overfitting Claudio F. Lima 1 Fernando G. Lobo 1 Martin Pelikan 2 1 Department of Electronics and Computer Science Engineering University of Algarve, Portugal 2

More information

CALCULATING MAGNETIC FIELDS & THE BIOT-SAVART LAW. Purdue University Physics 241 Lecture 15 Brendan Sullivan

CALCULATING MAGNETIC FIELDS & THE BIOT-SAVART LAW. Purdue University Physics 241 Lecture 15 Brendan Sullivan CALCULATING MAGNETIC FIELDS & THE BIOT-SAVAT LAW Purdue University Physics 41 Lecture 15 Brendan Sullivan Introduction Brendan Sullivan, PHYS89, sullivb@purdue.edu Office Hours: By Appointment Just stop

More information

CSCE 750 Final Exam Answer Key Wednesday December 7, 2005

CSCE 750 Final Exam Answer Key Wednesday December 7, 2005 CSCE 750 Final Exam Answer Key Wednesday December 7, 2005 Do all problems. Put your answers on blank paper or in a test booklet. There are 00 points total in the exam. You have 80 minutes. Please note

More information

Overview (Fall 2007) Machine Translation Part III. Roadmap for the Next Few Lectures. Phrase-Based Models. Learning phrases from alignments

Overview (Fall 2007) Machine Translation Part III. Roadmap for the Next Few Lectures. Phrase-Based Models. Learning phrases from alignments Overview Learning phrases from alignments 6.864 (Fall 2007) Machine Translation Part III A phrase-based model Decoding in phrase-based models (Thanks to Philipp Koehn for giving me slides from his EACL

More information

Oikos. Appendix 1 and 2. o20751

Oikos. Appendix 1 and 2. o20751 Oikos o20751 Rosindell, J. and Cornell, S. J. 2013. Universal scaling of species-abundance distributions across multiple scales. Oikos 122: 1101 1111. Appendix 1 and 2 Universal scaling of species-abundance

More information

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University SEQUENTIAL DATA So far, when thinking

More information

SUMMARIZING MEASURED DATA. Gaia Maselli

SUMMARIZING MEASURED DATA. Gaia Maselli SUMMARIZING MEASURED DATA Gaia Maselli maselli@di.uniroma1.it Computer Network Performance 2 Overview Basic concepts Summarizing measured data Summarizing data by a single number Summarizing variability

More information

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS PROBABILITY AND INFORMATION THEORY Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Probability space Rules of probability

More information

Proofs and derivations

Proofs and derivations A Proofs and derivations Proposition 1. In the sheltering decision problem of section 1.1, if m( b P M + s) = u(w b P M + s), where u( ) is weakly concave and twice continuously differentiable, then f

More information

Regression I: Mean Squared Error and Measuring Quality of Fit

Regression I: Mean Squared Error and Measuring Quality of Fit Regression I: Mean Squared Error and Measuring Quality of Fit -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD 1 The Setup Suppose there is a scientific problem we are interested in solving

More information

STATISTICS 1 REVISION NOTES

STATISTICS 1 REVISION NOTES STATISTICS 1 REVISION NOTES Statistical Model Representing and summarising Sample Data Key words: Quantitative Data This is data in NUMERICAL FORM such as shoe size, height etc. Qualitative Data This is

More information

Star-Structured High-Order Heterogeneous Data Co-clustering based on Consistent Information Theory

Star-Structured High-Order Heterogeneous Data Co-clustering based on Consistent Information Theory Star-Structured High-Order Heterogeneous Data Co-clustering based on Consistent Information Theory Bin Gao Tie-an Liu Wei-ing Ma Microsoft Research Asia 4F Sigma Center No. 49 hichun Road Beijing 00080

More information

Density Modeling and Clustering Using Dirichlet Diffusion Trees

Density Modeling and Clustering Using Dirichlet Diffusion Trees p. 1/3 Density Modeling and Clustering Using Dirichlet Diffusion Trees Radford M. Neal Bayesian Statistics 7, 2003, pp. 619-629. Presenter: Ivo D. Shterev p. 2/3 Outline Motivation. Data points generation.

More information

Second main application of Chernoff: analysis of load balancing. Already saw balls in bins example. oblivious algorithms only consider self packet.

Second main application of Chernoff: analysis of load balancing. Already saw balls in bins example. oblivious algorithms only consider self packet. Routing Second main application of Chernoff: analysis of load balancing. Already saw balls in bins example synchronous message passing bidirectional links, one message per step queues on links permutation

More information

Class XII_All India_Mathematics_Set-2 SECTION C. Question numbers 23 to 29 carry 6 marks each.

Class XII_All India_Mathematics_Set-2 SECTION C. Question numbers 23 to 29 carry 6 marks each. SECTION C Question numbers 3 to 9 carr 6 marks each. 3. Find the equation of the plane passing through the line of intersection of the planes r ˆi 3j ˆ 6 0 r 3i ˆ ˆ j k ˆ 0, whose perpendicular distance

More information

Math 20 Spring Discrete Probability. Midterm Exam

Math 20 Spring Discrete Probability. Midterm Exam Math 20 Spring 203 Discrete Probability Midterm Exam Thursday April 25, 5:00 7:00 PM Your name (please print): Instructions: This is a closed book, closed notes exam. Use of calculators is not permitted.

More information

Lecture 3 Probability Basics

Lecture 3 Probability Basics Lecture 3 Probability Basics Thais Paiva STA 111 - Summer 2013 Term II July 3, 2013 Lecture Plan 1 Definitions of probability 2 Rules of probability 3 Conditional probability What is Probability? Probability

More information

Analysis of Algorithms. Unit 5 - Intractable Problems

Analysis of Algorithms. Unit 5 - Intractable Problems Analysis of Algorithms Unit 5 - Intractable Problems 1 Intractable Problems Tractable Problems vs. Intractable Problems Polynomial Problems NP Problems NP Complete and NP Hard Problems 2 In this unit we

More information

Efficient Semiparametric Estimators via Modified Profile Likelihood in Frailty & Accelerated-Failure Models

Efficient Semiparametric Estimators via Modified Profile Likelihood in Frailty & Accelerated-Failure Models NIH Talk, September 03 Efficient Semiparametric Estimators via Modified Profile Likelihood in Frailty & Accelerated-Failure Models Eric Slud, Math Dept, Univ of Maryland Ongoing joint project with Ilia

More information

Machine Learning. Lecture 9: Learning Theory. Feng Li.

Machine Learning. Lecture 9: Learning Theory. Feng Li. Machine Learning Lecture 9: Learning Theory Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Why Learning Theory How can we tell

More information

WESTMORELAND COUNTY PUBLIC SCHOOLS Integrated Instructional Pacing Guide and Checklist Algebra II

WESTMORELAND COUNTY PUBLIC SCHOOLS Integrated Instructional Pacing Guide and Checklist Algebra II WESTMORELAND COUNTY PUBLIC SCHOOLS 201 2014 Integrated Instructional Pacing Guide and Checklist Algebra II (s) ESS Vocabulary AII.4 Absolute Value Equations and Inequalities Absolute Value Equations and

More information

Comparison of Methods for Computation of Ideal Circulation Distribution on the Propeller

Comparison of Methods for Computation of Ideal Circulation Distribution on the Propeller Comparison of Methods for Computation of Ideal Circulation Distribution on the Propeller Ing. Jan Klesa vedoucí práce: doc. Ing. Daniel Hanus, CSc. Abstrakt Ideální rozložení cirkulace na vrtuli je vypočteno

More information

Quantifying Stochastic Model Errors via Robust Optimization

Quantifying Stochastic Model Errors via Robust Optimization Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations

More information

Curve Fitting Re-visited, Bishop1.2.5

Curve Fitting Re-visited, Bishop1.2.5 Curve Fitting Re-visited, Bishop1.2.5 Maximum Likelihood Bishop 1.2.5 Model Likelihood differentiation p(t x, w, β) = Maximum Likelihood N N ( t n y(x n, w), β 1). (1.61) n=1 As we did in the case of the

More information

Časopis pro pěstování matematiky a fysiky

Časopis pro pěstování matematiky a fysiky Časopis pro pěstování matematiky a fysiky Norman Levinson Criteria for the limit-point case for second order linear differential operators Časopis pro pěstování matematiky a fysiky, Vol. 74 (1949), No.

More information

EE Power Gain and Amplifier Design 10/31/2017

EE Power Gain and Amplifier Design 10/31/2017 EE 40458 Power Gain and Amplifier Design 10/31/017 Motivation Brief recap: We ve studied matching networks (several types, how to design them, bandwidth, how they work, etc ) Studied network analysis techniques

More information

ECS171: Machine Learning

ECS171: Machine Learning ECS171: Machine Learning Lecture 6: Training versus Testing (LFD 2.1) Cho-Jui Hsieh UC Davis Jan 29, 2018 Preamble to the theory Training versus testing Out-of-sample error (generalization error): What

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Probability Models for Bayesian Recognition

Probability Models for Bayesian Recognition Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIAG / osig Second Semester 06/07 Lesson 9 0 arch 07 Probability odels for Bayesian Recognition Notation... Supervised Learning for Bayesian

More information

Revealed Preference 2011

Revealed Preference 2011 Revealed Preference 2011 Motivation: 1. up until now we have started with preference and then described behaviour 2. revealed preference works backwards - start with behaviour and describe preferences

More information

2.1.3 The Testing Problem and Neave s Step Method

2.1.3 The Testing Problem and Neave s Step Method we can guarantee (1) that the (unknown) true parameter vector θ t Θ is an interior point of Θ, and (2) that ρ θt (R) > 0 for any R 2 Q. These are two of Birch s regularity conditions that were critical

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

FINAL: CS 6375 (Machine Learning) Fall 2014

FINAL: CS 6375 (Machine Learning) Fall 2014 FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for

More information

Math Review Sheet, Fall 2008

Math Review Sheet, Fall 2008 1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the

More information

CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization

CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization Tim Roughgarden February 28, 2017 1 Preamble This lecture fulfills a promise made back in Lecture #1,

More information

Dover- Sherborn High School Mathematics Curriculum Probability and Statistics

Dover- Sherborn High School Mathematics Curriculum Probability and Statistics Mathematics Curriculum A. DESCRIPTION This is a full year courses designed to introduce students to the basic elements of statistics and probability. Emphasis is placed on understanding terminology and

More information

Chapter 5-2: Clustering

Chapter 5-2: Clustering Chapter 5-2: Clustering Jilles Vreeken Revision 1, November 20 th typo s fixed: dendrogram Revision 2, December 10 th clarified: we do consider a point x as a member of its own ε-neighborhood 12 Nov 2015

More information

1 Fields and vector spaces

1 Fields and vector spaces 1 Fields and vector spaces In this section we revise some algebraic preliminaries and establish notation. 1.1 Division rings and fields A division ring, or skew field, is a structure F with two binary

More information

SO x is a cubed root of t

SO x is a cubed root of t 7.6nth Roots 1) What do we know about x because of the following equation x 3 = t? All in one.docx SO x is a cubed root of t 2) Definition of nth root: 3) Study example 1 4) Now try the following problem

More information

Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron

Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron CS446: Machine Learning, Fall 2017 Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron Lecturer: Sanmi Koyejo Scribe: Ke Wang, Oct. 24th, 2017 Agenda Recap: SVM and Hinge loss, Representer

More information

Coding on Countably Infinite Alphabets

Coding on Countably Infinite Alphabets Coding on Countably Infinite Alphabets Non-parametric Information Theory Licence de droits d usage Outline Lossless Coding on infinite alphabets Source Coding Universal Coding Infinite Alphabets Enveloppe

More information

Describing Distributions with Numbers

Describing Distributions with Numbers Topic 2 We next look at quantitative data. Recall that in this case, these data can be subject to the operations of arithmetic. In particular, we can add or subtract observation values, we can sort them

More information

If something is repeated, look for what stays unchanged!

If something is repeated, look for what stays unchanged! LESSON 5: INVARIANTS Invariants are very useful in problems concerning algorithms involving steps that are repeated again and again (games, transformations). While these steps are repeated, the system

More information

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES

ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES ORDER STATISTICS, QUANTILES, AND SAMPLE QUANTILES 1. Order statistics Let X 1,...,X n be n real-valued observations. One can always arrangetheminordertogettheorder statisticsx (1) X (2) X (n). SinceX (k)

More information

Kullback-Leibler Designs

Kullback-Leibler Designs Kullback-Leibler Designs Astrid JOURDAN Jessica FRANCO Contents Contents Introduction Kullback-Leibler divergence Estimation by a Monte-Carlo method Design comparison Conclusion 2 Introduction Computer

More information

So far we discussed random number generators that need to have the maximum length period.

So far we discussed random number generators that need to have the maximum length period. So far we discussed random number generators that need to have the maximum length period. Even the increment series has the maximum length period yet it is by no means random How do we decide if a generator

More information

Intermediate Math Circles February 14, 2018 Contest Prep: Number Theory

Intermediate Math Circles February 14, 2018 Contest Prep: Number Theory Intermediate Math Circles February 14, 2018 Contest Prep: Number Theory Part 1: Prime Factorization A prime number is an integer greater than 1 whose only positive divisors are 1 and itself. An integer

More information

MAT 146. Semester Exam Part II 100 points (Part II: 50 points) Calculator Used Impact on Course Grade: approximately 30% Score

MAT 146. Semester Exam Part II 100 points (Part II: 50 points) Calculator Used Impact on Course Grade: approximately 30% Score MAT 146 Semester Exam Part II Name 100 points (Part II: 50 points) Calculator Used Impact on Course Grade: approximately 30% Score Questions (17) through (26) are each worth 5 points. See the grading rubric

More information

Algebra II Unit Breakdown (Curriculum Map Outline)

Algebra II Unit Breakdown (Curriculum Map Outline) QUARTER 1 Unit 1 (Arithmetic and Geometric Sequences) A. Sequences as Functions 1. Identify finite and infinite sequences 2. Identify an arithmetic sequence and its parts 3. Identify a geometric sequence

More information

Johns Hopkins University, Department of Mathematics Abstract Algebra - Spring 2013 Midterm Exam Solution

Johns Hopkins University, Department of Mathematics Abstract Algebra - Spring 2013 Midterm Exam Solution Johns Hopkins University, Department of Mathematics 110.40 Abstract Algebra - Spring 013 Midterm Exam Solution Instructions: This exam has 6 pages. No calculators, books or notes allowed. You must answer

More information

Dept of Computer Science, Michigan State University b

Dept of Computer Science, Michigan State University b CONTOUR REGRESSION: A distribution-regularized regression framework for climate modeling Zubin Abraham a, Pang-Ning Tan a, Julie A. Winkler b, Perdinan b, Shiyuan Zhong b, Malgorzata Liszewska c a Dept

More information

Indian Statistical Institute

Indian Statistical Institute Indian Statistical Institute Introductory Computer programming Robust Regression methods with high breakdown point Author: Roll No: MD1701 February 24, 2018 Contents 1 Introduction 2 2 Criteria for evaluating

More information

Predicting the Probability of Correct Classification

Predicting the Probability of Correct Classification Predicting the Probability of Correct Classification Gregory Z. Grudic Department of Computer Science University of Colorado, Boulder grudic@cs.colorado.edu Abstract We propose a formulation for binary

More information

CS570 Introduction to Data Mining

CS570 Introduction to Data Mining CS570 Introduction to Data Mining Department of Mathematics and Computer Science Li Xiong Data Exploration and Data Preprocessing Data and Attributes Data exploration Data pre-processing Data cleaning

More information

Vers un apprentissage subquadratique pour les mélanges d arbres

Vers un apprentissage subquadratique pour les mélanges d arbres Vers un apprentissage subquadratique pour les mélanges d arbres F. Schnitzler 1 P. Leray 2 L. Wehenkel 1 fschnitzler@ulg.ac.be 1 Université deliège 2 Université de Nantes 10 mai 2010 F. Schnitzler (ULG)

More information

Fall 2014 CMSC250/250H Midterm II

Fall 2014 CMSC250/250H Midterm II Fall 2014 CMSC250/250H Midterm II Circle Your Section! 0101 (10am: 3120, Ladan) 0102 (11am: 3120, Ladan) 0103 (Noon: 3120, Peter) 0201 (2pm: 3120, Yi) 0202 (10am: 1121, Vikas) 0203 (11am: 1121, Vikas)

More information

Pattern Popularity in 132-Avoiding Permutations

Pattern Popularity in 132-Avoiding Permutations Pattern Popularity in 132-Avoiding Permutations The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher Rudolph,

More information

Bias in RNA sequencing and what to do about it

Bias in RNA sequencing and what to do about it Bias in RNA sequencing and what to do about it Walter L. (Larry) Ruzzo Computer Science and Engineering Genome Sciences University of Washington Fred Hutchinson Cancer Research Center Seattle, WA, USA

More information

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2016 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

ENSEMBLE FLOOD INUNDATION FORECASTING: A CASE STUDY IN THE TIDAL DELAWARE RIVER

ENSEMBLE FLOOD INUNDATION FORECASTING: A CASE STUDY IN THE TIDAL DELAWARE RIVER ENSEMBLE FLOOD INUNDATION FORECASTING: A CASE STUDY IN THE TIDAL DELAWARE RIVER Michael Gomez & Alfonso Mejia Civil and Environmental Engineering Pennsylvania State University 10/12/2017 Mid-Atlantic Water

More information

(1) Sort all time observations from least to greatest, so that the j th and (j + 1) st observations are ordered by t j t j+1 for all j = 1,..., J.

(1) Sort all time observations from least to greatest, so that the j th and (j + 1) st observations are ordered by t j t j+1 for all j = 1,..., J. AFFIRMATIVE ACTION AND HUMAN CAPITAL INVESTMENT 8. ONLINE APPENDIX TO ACCOMPANY Affirmative Action and Human Capital Investment: Theory and Evidence from a Randomized Field Experiment, by CHRISTOPHER COTTON,

More information

QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS

QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS Parvathinathan Venkitasubramaniam, Gökhan Mergen, Lang Tong and Ananthram Swami ABSTRACT We study the problem of quantization for

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!

More information

FEEG6017 lecture: Akaike's information criterion; model reduction. Brendan Neville

FEEG6017 lecture: Akaike's information criterion; model reduction. Brendan Neville FEEG6017 lecture: Akaike's information criterion; model reduction Brendan Neville bjn1c13@ecs.soton.ac.uk Occam's razor William of Occam, 1288-1348. All else being equal, the simplest explanation is the

More information

Unit Two Descriptive Biostatistics. Dr Mahmoud Alhussami

Unit Two Descriptive Biostatistics. Dr Mahmoud Alhussami Unit Two Descriptive Biostatistics Dr Mahmoud Alhussami Descriptive Biostatistics The best way to work with data is to summarize and organize them. Numbers that have not been summarized and organized are

More information

MATHEMATICS (New Specification)

MATHEMATICS (New Specification) MS WELSH JOINT EDUCATION COMMITTEE. CYD-BWYLLGOR ADDYSG CYMRU General Certificate of Education Advanced Subsidiary/Advanced Tystysgrif Addysg Gyffredinol Uwch Gyfrannol/Uwch MARKING SCHEMES JANUARY 6 MATHEMATICS

More information

COMS 4705, Fall Machine Translation Part III

COMS 4705, Fall Machine Translation Part III COMS 4705, Fall 2011 Machine Translation Part III 1 Roadmap for the Next Few Lectures Lecture 1 (last time): IBM Models 1 and 2 Lecture 2 (today): phrase-based models Lecture 3: Syntax in statistical machine

More information

Unit 6 Study Guide: Equations. Section 6-1: One-Step Equations with Adding & Subtracting

Unit 6 Study Guide: Equations. Section 6-1: One-Step Equations with Adding & Subtracting Unit 6 Study Guide: Equations DUE DATE: A Day: Dec 18 th B Day: Dec 19 th Name Period Score / Section 6-1: One-Step Equations with Adding & Subtracting Textbook Reference: Page 437 Vocabulary: Equation

More information

A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie

A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie Computational Biology Program Memorial Sloan-Kettering Cancer Center http://cbio.mskcc.org/leslielab

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Bonding in solids The interaction of electrons in neighboring atoms of a solid serves the very important function of holding the crystal together.

Bonding in solids The interaction of electrons in neighboring atoms of a solid serves the very important function of holding the crystal together. Bonding in solids The interaction of electrons in neighboring atoms of a solid serves the very important function of holding the crystal together. For example Nacl In the Nacl lattice, each Na atom is

More information

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting

More information

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag Decision Trees Nicholas Ruozzi University of Texas at Dallas Based on the slides of Vibhav Gogate and David Sontag Supervised Learning Input: labelled training data i.e., data plus desired output Assumption:

More information

Pollution Sources Detection via Principal Component Analysis and Rotation

Pollution Sources Detection via Principal Component Analysis and Rotation Pollution Sources Detection via Principal Component Analysis and Rotation Vanessa Kuentz 1 in collaboration with : Marie Chavent 1 Hervé Guégan 2 Brigitte Patouille 1 Jérôme Saracco 1,3 1 IMB, Université

More information

MATH 120. Test 1 Spring, 2012 DO ALL ASSIGNED PROBLEMS. Things to particularly study

MATH 120. Test 1 Spring, 2012 DO ALL ASSIGNED PROBLEMS. Things to particularly study MATH 120 Test 1 Spring, 2012 DO ALL ASSIGNED PROBLEMS Things to particularly study 1) Critical Thinking Basic strategies Be able to solve using the basic strategies, such as finding patterns, questioning,

More information

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016 Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,

More information

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Ann Inst Stat Math (0) 64:359 37 DOI 0.007/s0463-00-036-3 Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk Paul Vos Qiang Wu Received: 3 June 009 / Revised:

More information

Intermediate Algebra with Applications

Intermediate Algebra with Applications Lakeshore Technical College 10-804-118 Intermediate Algebra with Applications Course Outcome Summary Course Information Alternate Title Description Total Credits 4 Total Hours 72 Pre/Corequisites Prerequisite

More information

Algebra 2 College Prep Curriculum Maps

Algebra 2 College Prep Curriculum Maps Algebra 2 College Prep Curriculum Maps Unit 1: Polynomial, Rational, and Radical Relationships Unit 2: Modeling With Functions Unit 3: Inferences and Conclusions from Data Unit 4: Trigonometric Functions

More information