Home Page. Title Page. Page 1 of c. Go Back. Full Screen. Close. Quit
|
|
- Natalie Lambert
- 5 years ago
- Views:
Transcription
1 ênúo (15-16 ïä)) wðï Page 1 of c
2 Introduction to Mathematical Statistics Fifth Edition Authors: Robert V. Hogg and Allen T. Craig Page 2 of 100
3 Page 3 of 100
4 Assistant: Zhang Juan (q ) URL: Page 4 of 100
5 Assignments Consulting References Page 5 of 100
6 References 1. [»: ênúoùâ. ŒÆ 2. + t, 9: ênúo. uà ŒÆ 3. FW: VÇØ ênúo. I EŒÆ 4. î è, 4D : VÇÚO. ŒÆ Page 6 of 100
7 5. ŸmÓ: VÇØ ênúo. uà ŒÆ 6. ÛÖµVÇØ ênúo. p 7. R ½SPLUS Šó Page 7 of 100
8 þ žm /:µ ±þìµ10:40 12:10 ±oþìµ8:50 10:20 o µ321 Page 8 of 100
9 1 Page 9 of 100
10 Page 10 of 100
11 Page 11 of 100
12 Page 12 of 100
13 Page 13 of 100
14 Page 14 of 100
15 Page 15 of 100
16 2 Page 16 of 100
17 êæ <hl ÚO4<²œ <) œ, Page 17 of 100 žæðúoœ
18 ež)þ23žÿo Ò7L ÆÚO Ï ÚOÆÒ 3ÿþ " )Florence Nightingale Page 18 of 100
19 ÚOÆ Statistics Â8 ÛÚ?nêâ ÆÚ²â Page 19 of 100 5Ø6z Ö6
20 { u y. Statisticum Collegium ( I )! Œ Statista ( I ½ [) ±9 cstatistik (?Ú O) L«ïÄI[ Æ" Page 20 of 100
21 ÚOÆ) uð Ú)åu Ð!?Ú; ƒé" ÚOÆ)µ % í # 11676c J Ñ / Ž â 0 ² L ÿ Ý # K A u1662cjñ< CÄÿÝ { I " Page 21 of 100
22 y ÚOuÐmൠ19 V",ŒÆm/IœV 0½/ Žâ0 Åì ž, ƒ /ÚO Û Æ0 " Page 22 of 100
23 ÚOÆÄ:/ žïµ 18 V" 19 V"/ ±ênú O Ä:ĵe".Ê.du1802c3IÚOÅ mð²l N¹Ä JÑÄ NVg uðƒ'eâ" Page 23 of 100
24 1805cV4uÐ {" 1809cpdêÆ[ÅìïáØ ÙnØ C½y ÚO ÏnØÄ:" ' žpavu1835c 1846cm òvçø %4 ½n ÙnØÚ\ ²LïÄ" Page 24 of 100
25 1870c puy 8 ƒ'v g I XÚOíäž 5" ù@ïóš ÚOÆïá Äuêâ½,5Aïĵe 3ù žï/ ênúoæú ÚOÆ" Page 25 of 100
26 20 V±5 ÆEâ ˆuÐ u)ãœcz ÚOÆ?\nØN XzuÐ ÙžÏ" k# Öu1900cJÑ[Ü`Ý u x* y Æb`ƒm ål"< UŠâ*ÿµdb` Ün5" Page 26 of 100
27 1908cxÚAJÑt ÙVg9 nø I XëêOnØÄ: µe " u cmjñwí 5u,uÐ ÛnØÚÁ OnØ" Page 27 of 100
28 1930c GùÚOŠ# Ö `unø ÚK ÚOÆu Ð (áäuî êæü6ï ÚOÆnØNXuÐ " 20 %ÇuÐ Ÿ Ò" Page 28 of 100
29 Œ 7 ± ê â Ø % & ê â 5 Æ A! ' X Ú C z 9 S A ^ 8 I y ÚOÆ " {Ø ÆNXÅì/ Page 29 of 100
30 êâ &E1N knõa.µ ½5!½þ!"!Ø!p! þ!ã/... Page 30 of 100
31 Ch. 6. Introduction to Statistical Inference 6.1. Point Estimation The first five chapters: Page 31 of 100
32 1. Concepts and probability theory. Random variable (r.v.) Distribution of r.v. Characteristic Moment Page 32 of 100
33 2. Statistical distribution. Discrete: Binomial, Poisson, Parskal, Discrete-uniform, Geometric, Hypergeometric, Multinomial, etc. Page 33 of 100
34 Continuous: Normal, Uniform, χ 2, t, F, Gamma, Exponential, Cauchy, Beta, Lognormal, Logistic, Multinormal, etc. Page 34 of 100
35 Page 35 of 100
36 Page 36 of 100
37 Page 37 of 100
38 Page 38 of 100
39 Page 39 of 100
40 Page 40 of 100
41 Page 41 of 100
42 Page 42 of 100
43 Page 43 of 100
44 Page 44 of 100
45 D.f. and c.d.f. Random representation Expectation and variance Median and MAD Skewness and kurtosis R or SPLUS language (data, programme, picture, etc.) Page 45 of 100
46 Convergence of random variable sequence The Law of Large Numbers Central Limit Theorem Page 46 of 100
47 A study of some problems in statistics A sample space C of outcomes and the space A of one or more random variables defined on C. A random variable X as the outcome of a random experiment Page 47 of 100
48 Call X the numerical outcome. Once the experiment has been performed and it is found that X = x, we shall call x the experimental value of X. Page 48 of 100
49 Let a random experiment be repeated n independent times and under identical conditions. Page 49 of 100
50 Random variables (a random sample) X 1, X 2,, X n with the observations. It is found that X 1 = x 1, X 2 = x 2,, X n = x n, we shall refer to x 1, x 2,, x n as the experimental values of X 1, X 2,, X n or as the sample data. Page 50 of 100
51 Let a r.v. X have a p.d.f. that is of known functional form but depends upon an unknown parameter θ: {f(x; θ), θ Ω} Ω: parameter space. Page 51 of 100
52 Not with one, but with a family of distributions. To each value of θ, θ Ω, there corresponds one member of the family. Page 52 of 100
53 Normal distribution family: {N(θ, 1) : θ Ω}, where Ω is the set < θ <. Binomial distribution family: {B(n, p) : p Ω}, where Ω is the set 0 p 1. Page 53 of 100
54 Need to select precisely one member of the family as being the p.d.f. of his random variable: point estimate of θ. Page 54 of 100
55 Let X 1, X 2,, X n denote a random sample from the family {f(x; θ), θ Ω}. Define a statistic Y 1 = u 1 (X 1, X 2,, X n ) as an estimate of θ. Page 55 of 100
56 If x 1, x 2,, x n are the observed experimental values of X 1, X 2,, X n, then the number y 1 = u 1 (x 1, x 2,, x n ) will be a good point estimate of θ. Page 56 of 100
57 It should not depend on θ. One principle that is often used in finding point estimates. Page 57 of 100
58 Example 1. Let X 1, X 2,, X n denote a random sample from the distribution with p.d.f. f(x) = where 0 θ 1. θ x (1 θ) (1 x) x = 0, 1 0 elsewhere Page 58 of 100
59 The probability that X 1 = x 1, X 2 = x 2,, X n = x n is the joint p.d.f. θ x 1 (1 θ) (1 x 1) θ x 2 (1 θ) 1 x 2 θ x n (1 θ) 1 x n = θ x i (1 θ) n x i Page 59 of 100 where x i equals zero or 1, i = 1, 2,, n.
60 This probability,which is the joint p.d.f. of X 1, X 2,, X n, may be regarded as a function of θ and denoted it by L(θ) and called the likelihood function. That is, L(θ) = θ x i (1 θ) n x i Page 60 of θ 1.
61 What value of θ would maximize the probability L(θ) of obtaining this particular sample x 1, x 2,, x n? Page 61 of 100
62 The maximizing value of θ would seemingly be a good estimate of θ because it would provide the largest probability of this particular sample. Page 62 of 100
63 Either L(θ) or ln L(θ) can be used, since the likelihood function L(θ) and its logarithm, ln L(θ), are maximized for the same value θ. Page 63 of 100
64 Here ln L(θ) = ( n x i ) ln θ+(n n x i ) ln(1 θ); we have d ln L(θ) dθ 1 = xi θ provided that θ 0, 1. 1 n x i 1 θ = 0 Page 64 of 100
65 This is equivalent to the equation n n (1 θ) x i = θ(n x i ). 1 The solution for θ is n 1 x i/n, it maximizes L(θ) and ln L(θ) can be easily checked ( even for all of x 1, x 2,, x n = 0 or 1). 1 Page 65 of 100
66 Then n 1 x i/n is the value of θ that maximizes L(θ). The corresponding statistic, ˆθ = 1 n X i = n X i=1 is called the maximum likelihood estimator (MLE) of θ. Page 66 of 100
67 The observed value of ˆθ, namely n 1 x i/n, is called the value of maximum likelihood estimator (MLE) of θ. For example, suppose that n = 3, and x 1 = 1, x 2 = 0, x 3 = 1, then L(θ) = θ 2 (1 θ) and the observed ˆθ = 2 3 is the MLE of θ. Page 67 of 100
68 The principle of maximum likelihood: Consider a random sample X 1, X 2,, X n f(x, θ), θ Ω. The joint p.d.f. of X 1, X 2,, X n is f(x 1 ; θ)f(x 2 ; θ) f(x n ; θ). Page 68 of 100
69 This joint p.d.f. may be regarded as a function of θ. It is called the likelihood function L of the random sample. Write for θ Ω, L(θ; x 1, x 2,, x n ) = f(x 1 ; θ)f(x 2 ; θ) f(x n ; θ). Page 69 of 100
70 Suppose that u(x 1, x 2,, x n ) is a nontrivial function such that, when θ is replaced by u(x 1, x 2,, x n ), the likelihood function L is maximized: L[u(x 1, x 2,, x n ); x 1, x 2,, x n ] = max θ Ω L(θ; x 1, x 2,, x n ). Page 70 of 100
71 Then the statistic u(x 1, x 2,, x n ) will be called a maximum likelihood estimator(mle) of θ and will be denoted by the symbol ˆθ = u(x 1, x 2,, x n ). Page 71 of 100
72 In many instances there will be a unique MLE ˆθ of a parameter θ, It may be obtained by the process of differentiation. Page 72 of 100
73 Example 2. Let X 1, X 2,, X n N(θ, 1), < θ <. Here L(θ; x 1, x 2,, x n ) 1 n = ( ) n (x i θ) 2 exp[ ] 2π 2 1 Page 73 of 100
74 This function L can be maximized by setting the first derivative of L, with respect to θ for ln L. It may be easier to solve d ln L(θ; x 1, x 2,, x n ) dθ = 0. Page 74 of 100
75 d ln L(θ; x 1, x 2,, x n ) dθ = n (x i θ) = 0, the solution for the parameter θ is n u(x 1, x 2,, x n ) = x i /n. That n 1 x i/n actually maximizes L. 1 1 Page 75 of 100
76 Thus the statistic ˆθ = u(x 1, X 2,, X n ) = 1 n is the unique MLE of the mean θ. n X i = X 1 Page 76 of 100
77 From Examples 1, 2, E(ˆθ) = θ. The expected value of the estimator is equal to the corresponding parameter. Page 77 of 100
78 Definition: Any statistic whose mathematical expectation is equal to a parameter θ is called an unbiased estimator of the parameter θ. Otherwise, the statistic is said to be biased. Page 78 of 100
79 Example 3. Let 1 θ 0 < x θ, 0 < θ <, f(x; θ) = 0 elsewhere X 1, X 2,, X n denote a random sample from f(x; θ). Note that 0 < x θ. Page 79 of 100
80 We have L(θ; x 1, x 2,, x n ) = 1 θ n, 0 < x i θ Contents which is an ever-decreasing function of θ. The maximum of such functions cannot be found by differentiation. Page 80 of 100
81 L can be made no larger than 1 [max(x i )] n and the unique MLE ˆθ of θ is the nth order statistic max(x i ). E[max(X i )] = nθ/(n + 1). The MLE of the parameter θ is biased. Page 81 of 100
Probability Distributions Columns (a) through (d)
Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)
More informationThree hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.
Three hours To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER EXTREME VALUES AND FINANCIAL RISK Examiner: Answer QUESTION 1, QUESTION
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationChapter 8.8.1: A factorization theorem
LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More information[Chapter 6. Functions of Random Variables]
[Chapter 6. Functions of Random Variables] 6.1 Introduction 6.2 Finding the probability distribution of a function of random variables 6.3 The method of distribution functions 6.5 The method of Moment-generating
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationSTA 260: Statistics and Probability II
Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition
More informationStatistics Ph.D. Qualifying Exam: Part I October 18, 2003
Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More information! " # $! % & '! , ) ( + - (. ) ( ) * + / 0 1 2 3 0 / 4 5 / 6 0 ; 8 7 < = 7 > 8 7 8 9 : Œ Š ž P P h ˆ Š ˆ Œ ˆ Š ˆ Ž Ž Ý Ü Ý Ü Ý Ž Ý ê ç è ± ¹ ¼ ¹ ä ± ¹ w ç ¹ è ¼ è Œ ¹ ± ¹ è ¹ è ä ç w ¹ ã ¼ ¹ ä ¹ ¼ ¹ ±
More informationTest Problems for Probability Theory ,
1 Test Problems for Probability Theory 01-06-16, 010-1-14 1. Write down the following probability density functions and compute their moment generating functions. (a) Binomial distribution with mean 30
More informationFramework for functional tree simulation applied to 'golden delicious' apple trees
Purdue University Purdue e-pubs Open Access Theses Theses and Dissertations Spring 2015 Framework for functional tree simulation applied to 'golden delicious' apple trees Marek Fiser Purdue University
More informationSTATISTICS SYLLABUS UNIT I
STATISTICS SYLLABUS UNIT I (Probability Theory) Definition Classical and axiomatic approaches.laws of total and compound probability, conditional probability, Bayes Theorem. Random variable and its distribution
More informationAn Example file... log.txt
# ' ' Start of fie & %$ " 1 - : 5? ;., B - ( * * B - ( * * F I / 0. )- +, * ( ) 8 8 7 /. 6 )- +, 5 5 3 2( 7 7 +, 6 6 9( 3 5( ) 7-0 +, => - +< ( ) )- +, 7 / +, 5 9 (. 6 )- 0 * D>. C )- +, (A :, C 0 )- +,
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationCourse: ESO-209 Home Work: 1 Instructor: Debasis Kundu
Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear
More informationStatistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm
Statistics GIDP Ph.D. Qualifying Exam Theory Jan, 06, 9:00am-:00pm Instructions: Provide answers on the supplied pads of paper; write on only one side of each sheet. Complete exactly 5 of the 6 problems.
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationf(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain
0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationHT Introduction. P(X i = x i ) = e λ λ x i
MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More informationTheory of Statistical Tests
Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationMcGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper
McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section
More informationHypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationPractice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:
Practice Exam 1 1. Losses for an insurance coverage have the following cumulative distribution function: F(0) = 0 F(1,000) = 0.2 F(5,000) = 0.4 F(10,000) = 0.9 F(100,000) = 1 with linear interpolation
More informationGeneralized Linear Models Introduction
Generalized Linear Models Introduction Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Linear Models For many problems, standard linear regression approaches don t work. Sometimes,
More informationOC330C. Wiring Diagram. Recommended PKH- P35 / P50 GALH PKA- RP35 / RP50. Remarks (Drawing No.) No. Parts No. Parts Name Specifications
G G " # $ % & " ' ( ) $ * " # $ % & " ( + ) $ * " # C % " ' ( ) $ * C " # C % " ( + ) $ * C D ; E @ F @ 9 = H I J ; @ = : @ A > B ; : K 9 L 9 M N O D K P D N O Q P D R S > T ; U V > = : W X Y J > E ; Z
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationIE 303 Discrete-Event Simulation
IE 303 Discrete-Event Simulation 1 L E C T U R E 5 : P R O B A B I L I T Y R E V I E W Review of the Last Lecture Random Variables Probability Density (Mass) Functions Cumulative Density Function Discrete
More informationReview for the previous lecture
Lecture 1 and 13 on BST 631: Statistical Theory I Kui Zhang, 09/8/006 Review for the previous lecture Definition: Several discrete distributions, including discrete uniform, hypergeometric, Bernoulli,
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationLA PRISE DE CALAIS. çoys, çoys, har - dis. çoys, dis. tons, mantz, tons, Gas. c est. à ce. C est à ce. coup, c est à ce
> ƒ? @ Z [ \ _ ' µ `. l 1 2 3 z Æ Ñ 6 = Ð l sl (~131 1606) rn % & +, l r s s, r 7 nr ss r r s s s, r s, r! " # $ s s ( ) r * s, / 0 s, r 4 r r 9;: < 10 r mnz, rz, r ns, 1 s ; j;k ns, q r s { } ~ l r mnz,
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationTwo hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.
Two hours MATH38181 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER EXTREME VALUES AND FINANCIAL RISK Examiner: Answer any FOUR
More informationCOPYRIGHTED MATERIAL CONTENTS. Preface Preface to the First Edition
Preface Preface to the First Edition xi xiii 1 Basic Probability Theory 1 1.1 Introduction 1 1.2 Sample Spaces and Events 3 1.3 The Axioms of Probability 7 1.4 Finite Sample Spaces and Combinatorics 15
More informationDeccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY. SECOND YEAR B.Sc. SEMESTER - III
Deccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY SECOND YEAR B.Sc. SEMESTER - III SYLLABUS FOR S. Y. B. Sc. STATISTICS Academic Year 07-8 S.Y. B.Sc. (Statistics)
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationPlanning for Reactive Behaviors in Hide and Seek
University of Pennsylvania ScholarlyCommons Center for Human Modeling and Simulation Department of Computer & Information Science May 1995 Planning for Reactive Behaviors in Hide and Seek Michael B. Moore
More informationTwo hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45
Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved
More informationBMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution
Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random
More informationprobability of k samples out of J fall in R.
Nonparametric Techniques for Density Estimation (DHS Ch. 4) n Introduction n Estimation Procedure n Parzen Window Estimation n Parzen Window Example n K n -Nearest Neighbor Estimation Introduction Suppose
More informationDepartment of Statistical Science FIRST YEAR EXAM - SPRING 2017
Department of Statistical Science Duke University FIRST YEAR EXAM - SPRING 017 Monday May 8th 017, 9:00 AM 1:00 PM NOTES: PLEASE READ CAREFULLY BEFORE BEGINNING EXAM! 1. Do not write solutions on the exam;
More informationPractice Problems Section Problems
Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,
More informationIt's Only Fitting. Fitting model to data parameterizing model estimating unknown parameters in the model
It's Only Fitting Fitting model to data parameterizing model estimating unknown parameters in the model Likelihood: an example Cohort of 8! individuals observe survivors at times >œ 1, 2, 3,..., : 8",
More informationGradient Ascent Chris Piech CS109, Stanford University
Gradient Ascent Chris Piech CS109, Stanford University Our Path Deep Learning Linear Regression Naïve Bayes Logistic Regression Parameter Estimation Our Path Deep Learning Linear Regression Naïve Bayes
More informationLecture 2. Distributions and Random Variables
Lecture 2. Distributions and Random Variables Igor Rychlik Chalmers Department of Mathematical Sciences Probability, Statistics and Risk, MVE300 Chalmers March 2013. Click on red text for extra material.
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationConstructive Decision Theory
Constructive Decision Theory Joe Halpern Cornell University Joint work with Larry Blume and David Easley Economics Cornell Constructive Decision Theory p. 1/2 Savage s Approach Savage s approach to decision
More informationChing-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12
Lecture 5 Continuous Random Variables BMIR Lecture Series in Probability and Statistics Ching-Han Hsu, BMES, National Tsing Hua University c 215 by Ching-Han Hsu, Ph.D., BMIR Lab 5.1 1 Uniform Distribution
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More informationDS-GA 1002 Lecture notes 11 Fall Bayesian statistics
DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian
More informationt x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.
Mathematical Statistics: Homewor problems General guideline. While woring outside the classroom, use any help you want, including people, computer algebra systems, Internet, and solution manuals, but mae
More informationStatistics 1B. Statistics 1B 1 (1 1)
0. Statistics 1B Statistics 1B 1 (1 1) 0. Lecture 1. Introduction and probability review Lecture 1. Introduction and probability review 2 (1 1) 1. Introduction and probability review 1.1. What is Statistics?
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationChapter 3 Common Families of Distributions
Lecture 9 on BST 631: Statistical Theory I Kui Zhang, 9/3/8 and 9/5/8 Review for the previous lecture Definition: Several commonly used discrete distributions, including discrete uniform, hypergeometric,
More informationRedoing the Foundations of Decision Theory
Redoing the Foundations of Decision Theory Joe Halpern Cornell University Joint work with Larry Blume and David Easley Economics Cornell Redoing the Foundations of Decision Theory p. 1/21 Decision Making:
More informationReview of Probabilities and Basic Statistics
Alex Smola Barnabas Poczos TA: Ina Fiterau 4 th year PhD student MLD Review of Probabilities and Basic Statistics 10-701 Recitations 1/25/2013 Recitation 1: Statistics Intro 1 Overview Introduction to
More informationSTATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero
STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank
More informationParameter Estimation
Parameter Estimation Chapters 13-15 Stat 477 - Loss Models Chapters 13-15 (Stat 477) Parameter Estimation Brian Hartman - BYU 1 / 23 Methods for parameter estimation Methods for parameter estimation Methods
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationLikelihoods. P (Y = y) = f(y). For example, suppose Y has a geometric distribution on 1, 2,... with parameter p. Then the pmf is
Likelihoods The distribution of a random variable Y with a discrete sample space (e.g. a finite sample space or the integers) can be characterized by its probability mass function (pmf): P (Y = y) = f(y).
More informationNew Bayesian methods for model comparison
Back to the future New Bayesian methods for model comparison Murray Aitkin murray.aitkin@unimelb.edu.au Department of Mathematics and Statistics The University of Melbourne Australia Bayesian Model Comparison
More informationSuggested solutions to written exam Jan 17, 2012
LINKÖPINGS UNIVERSITET Institutionen för datavetenskap Statistik, ANd 73A36 THEORY OF STATISTICS, 6 CDTS Master s program in Statistics and Data Mining Fall semester Written exam Suggested solutions to
More informationSTAB57: Quiz-1 Tutorial 1 (Show your work clearly) 1. random variable X has a continuous distribution for which the p.d.f.
STAB57: Quiz-1 Tutorial 1 1. random variable X has a continuous distribution for which the p.d.f. is as follows: { kx 2.5 0 < x < 1 f(x) = 0 otherwise where k > 0 is a constant. (a) (4 points) Determine
More informationLecture 4 September 15
IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric
More informationProblem 1 (20) Log-normal. f(x) Cauchy
ORF 245. Rigollet Date: 11/21/2008 Problem 1 (20) f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.0 0.2 0.4 0.6 0.8 4 2 0 2 4 Normal (with mean -1) 4 2 0 2 4 Negative-exponential x x f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.5
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More information1. Fisher Information
1. Fisher Information Let f(x θ) be a density function with the property that log f(x θ) is differentiable in θ throughout the open p-dimensional parameter set Θ R p ; then the score statistic (or score
More informationChapters 9. Properties of Point Estimators
Chapters 9. Properties of Point Estimators Recap Target parameter, or population parameter θ. Population distribution f(x; θ). { probability function, discrete case f(x; θ) = density, continuous case The
More information1. Point Estimators, Review
AMS571 Prof. Wei Zhu 1. Point Estimators, Review Example 1. Let be a random sample from. Please find a good point estimator for Solutions. There are the typical estimators for and. Both are unbiased estimators.
More informationDistribution Fitting (Censored Data)
Distribution Fitting (Censored Data) Summary... 1 Data Input... 2 Analysis Summary... 3 Analysis Options... 4 Goodness-of-Fit Tests... 6 Frequency Histogram... 8 Comparison of Alternative Distributions...
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More information15 Discrete Distributions
Lecture Note 6 Special Distributions (Discrete and Continuous) MIT 4.30 Spring 006 Herman Bennett 5 Discrete Distributions We have already seen the binomial distribution and the uniform distribution. 5.
More informationModern Methods of Data Analysis - WS 07/08
Modern Methods of Data Analysis Lecture VIc (19.11.07) Contents: Maximum Likelihood Fit Maximum Likelihood (I) Assume N measurements of a random variable Assume them to be independent and distributed according
More informationChapter Learning Objectives. Probability Distributions and Probability Density Functions. Continuous Random Variables
Chapter 4: Continuous Random Variables and Probability s 4-1 Continuous Random Variables 4-2 Probability s and Probability Density Functions 4-3 Cumulative Functions 4-4 Mean and Variance of a Continuous
More informationFinal Examination a. STA 532: Statistical Inference. Wednesday, 2015 Apr 29, 7:00 10:00pm. Thisisaclosed bookexam books&phonesonthefloor.
Final Examination a STA 532: Statistical Inference Wednesday, 2015 Apr 29, 7:00 10:00pm Thisisaclosed bookexam books&phonesonthefloor Youmayuseacalculatorandtwo pagesofyourownnotes Do not share calculators
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationECON 4160, Autumn term Lecture 1
ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationChapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics
Chapter 6 Order Statistics and Quantiles 61 Extreme Order Statistics Suppose we have a finite sample X 1,, X n Conditional on this sample, we define the values X 1),, X n) to be a permutation of X 1,,
More informationChapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.
Chapter 2 Random Variable CLO2 Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance. 1 1. Introduction In Chapter 1, we introduced the concept
More informationMATH4427 Notebook 2 Fall Semester 2017/2018
MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationAn Introduction to Optimal Control Applied to Disease Models
An Introduction to Optimal Control Applied to Disease Models Suzanne Lenhart University of Tennessee, Knoxville Departments of Mathematics Lecture1 p.1/37 Example Number of cancer cells at time (exponential
More information6 The normal distribution, the central limit theorem and random samples
6 The normal distribution, the central limit theorem and random samples 6.1 The normal distribution We mentioned the normal (or Gaussian) distribution in Chapter 4. It has density f X (x) = 1 σ 1 2π e
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationBrief Review of Probability
Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions
More informationSummary of Chapters 7-9
Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationParameter Estimation
Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y
More informationContinuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( ) Chapter 4 4.
UCLA STAT 11 A Applied Probability & Statistics for Engineers Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology Teaching Assistant: Christopher Barr University of California, Los Angeles,
More informationMTH 452 Mathematical Statistics
MTH 452 Mathematical Statistics Instructor: Orlando Merino University of Rhode Island Spring Semester, 2006 1 5.1 Introduction An Experiment: In 10 consecutive trips to the free throw line, a professional
More informationStatistical Methods in HYDROLOGY CHARLES T. HAAN. The Iowa State University Press / Ames
Statistical Methods in HYDROLOGY CHARLES T. HAAN The Iowa State University Press / Ames Univariate BASIC Table of Contents PREFACE xiii ACKNOWLEDGEMENTS xv 1 INTRODUCTION 1 2 PROBABILITY AND PROBABILITY
More information