MISS DISTANCE GENERALIZED VARIANCE NON-CENTRAL CHI DISTRIBUTION. Ken Chan ABSTRACT

Similar documents
10. Joint Moments and Joint Characteristic Functions

Estimation and detection of a periodic signal

Physics 5153 Classical Mechanics. Solution by Quadrature-1

ELEG 3143 Probability & Stochastic Process Ch. 4 Multiple Random Variables

Review of Prerequisite Skills for Unit # 2 (Derivatives) U2L2: Sec.2.1 The Derivative Function

Numerical Methods - Lecture 2. Numerical Methods. Lecture 2. Analysis of errors in numerical methods

Basic properties of limits

3. Several Random Variables

Special types of Riemann sums

Mathematical Notation Math Calculus & Analytic Geometry III

. This is the Basic Chain Rule. x dt y dt z dt Chain Rule in this context.

Feedback Linearization

MODULE 6 LECTURE NOTES 1 REVIEW OF PROBABILITY THEORY. Most water resources decision problems face the risk of uncertainty mainly because of the

Objectives. By the time the student is finished with this section of the workbook, he/she should be able

The achievable limits of operational modal analysis. * Siu-Kui Au 1)

Chapter 6 Reliability-based design and code developments

Roberto s Notes on Differential Calculus Chapter 8: Graphical analysis Section 1. Extreme points

Least-Squares Spectral Analysis Theory Summary

Curve Sketching. The process of curve sketching can be performed in the following steps:

and ( x, y) in a domain D R a unique real number denoted x y and b) = x y = {(, ) + 36} that is all points inside and on

1. Definition: Order Statistics of a sample.

2.6 Two-dimensional continuous interpolation 3: Kriging - introduction to geostatistics. References - geostatistics. References geostatistics (cntd.

Mathematical Notation Math Calculus & Analytic Geometry III

Optimal Control. with. Aerospace Applications. James M. Longuski. Jose J. Guzman. John E. Prussing

Mat 267 Engineering Calculus III Updated on 9/19/2010

PHY 335 Data Analysis for Physicists

CHAPTER 1: INTRODUCTION. 1.1 Inverse Theory: What It Is and What It Does

APPENDIX 1 ERROR ESTIMATION

2. ETA EVALUATIONS USING WEBER FUNCTIONS. Introduction

A Simple Explanation of the Sobolev Gradient Method

( x) f = where P and Q are polynomials.

DETC A GENERALIZED MAX-MIN SAMPLE FOR RELIABILITY ASSESSMENT WITH DEPENDENT VARIABLES

Answer Key-Math 11- Optional Review Homework For Exam 2

Probabilistic Optimisation applied to Spacecraft Rendezvous on Keplerian Orbits

Products and Convolutions of Gaussian Probability Density Functions

In many diverse fields physical data is collected or analysed as Fourier components.

Categories and Natural Transformations

Ex x xf xdx. Ex+ a = x+ a f x dx= xf x dx+ a f xdx= xˆ. E H x H x H x f x dx ˆ ( ) ( ) ( ) μ is actually the first moment of the random ( )

Chapter 2. Basic concepts of probability. Summary. 2.1 Axiomatic foundation of probability theory

Additional exercises in Stationary Stochastic Processes

8. INTRODUCTION TO STATISTICAL THERMODYNAMICS

1 Relative degree and local normal forms

The Deutsch-Jozsa Problem: De-quantization and entanglement

AP Calculus Notes: Unit 1 Limits & Continuity. Syllabus Objective: 1.1 The student will calculate limits using the basic limit theorems.

Estimation of Sample Reactivity Worth with Differential Operator Sampling Method

Definition: Let f(x) be a function of one variable with continuous derivatives of all orders at a the point x 0, then the series.

Gaussian Process Regression Models for Predicting Stock Trends

Analog Computing Technique

RELIABILITY OF BURIED PIPELINES WITH CORROSION DEFECTS UNDER VARYING BOUNDARY CONDITIONS

CHAPTER 8 ANALYSIS OF AVERAGE SQUARED DIFFERENCE SURFACES

Bayesian Technique for Reducing Uncertainty in Fatigue Failure Model

Differentiation. The main problem of differential calculus deals with finding the slope of the tangent line at a point on a curve.

Example: When describing where a function is increasing, decreasing or constant we use the x- axis values.

Power Spectral Analysis of Elementary Cellular Automata

Chapter 11 Collision Theory

NONLINEAR CONTROL OF POWER NETWORK MODELS USING FEEDBACK LINEARIZATION

Telescoping Decomposition Method for Solving First Order Nonlinear Differential Equations

TLT-5200/5206 COMMUNICATION THEORY, Exercise 3, Fall TLT-5200/5206 COMMUNICATION THEORY, Exercise 3, Fall Problem 1.

Scattered Data Approximation of Noisy Data via Iterated Moving Least Squares

SIO 211B, Rudnick. We start with a definition of the Fourier transform! ĝ f of a time series! ( )

Introduction to Simulation - Lecture 2. Equation Formulation Methods. Jacob White. Thanks to Deepak Ramaswamy, Michal Rewienski, and Karen Veroy

Extreme Values of Functions

Fluctuationlessness Theorem and its Application to Boundary Value Problems of ODEs

The Ascent Trajectory Optimization of Two-Stage-To-Orbit Aerospace Plane Based on Pseudospectral Method

Review D: Potential Energy and the Conservation of Mechanical Energy

9.3 Graphing Functions by Plotting Points, The Domain and Range of Functions

Wind-Driven Circulation: Stommel s gyre & Sverdrup s balance

Multi Variable Calculus

Chapter 8: MULTIPLE CONTINUOUS RANDOM VARIABLES

Math Review and Lessons in Calculus

Module 27: Rigid Body Dynamics: Rotation and Translation about a Fixed Axis

Physics 2B Chapter 17 Notes - First Law of Thermo Spring 2018

AH 2700A. Attenuator Pair Ratio for C vs Frequency. Option-E 50 Hz-20 khz Ultra-precision Capacitance/Loss Bridge

The concept of limit

On High-Rate Cryptographic Compression Functions

RESOLUTION MSC.362(92) (Adopted on 14 June 2013) REVISED RECOMMENDATION ON A STANDARD METHOD FOR EVALUATING CROSS-FLOODING ARRANGEMENTS

Multivariate Distribution Models

STAT 801: Mathematical Statistics. Hypothesis Testing

International Space Station Collision Probability

9.1 The Square Root Function

The Space of Negative Scalar Curvature Metrics

y,z the subscript y, z indicating that the variables y and z are kept constant. The second partial differential with respect to x is written x 2 y,z

UNIVERSITY OF CALIFORNIA - SANTA CRUZ DEPARTMENT OF PHYSICS PHYS 112. Homework #4. Benjamin Stahl. February 2, 2015

Lecture : Feedback Linearization

Asymptotics of integrals

Solutions for Homework #8. Landing gear

Continuity and Differentiability

On a Closed Formula for the Derivatives of e f(x) and Related Financial Applications

Chapter 4 Imaging. Lecture 21. d (110) Chem 793, Fall 2011, L. Ma

Lecture 25: Heat and The 1st Law of Thermodynamics Prof. WAN, Xin

Symbolic-Numeric Methods for Improving Structural Analysis of DAEs

Minimum-Time Trajectory Optimization of Multiple Revolution Low-Thrust Earth-Orbit Transfers

Syllabus Objective: 2.9 The student will sketch the graph of a polynomial, radical, or rational function.

Digital Image Processing. Lecture 6 (Enhancement) Bu-Ali Sina University Computer Engineering Dep. Fall 2009

Semideterministic Finite Automata in Operational Research

Joint ] X 5) P[ 6) P[ (, ) = y 2. x 1. , y. , ( x, y ) 2, (

(One Dimension) Problem: for a function f(x), find x 0 such that f(x 0 ) = 0. f(x)

( ) x y z. 3 Functions 36. SECTION D Composite Functions

An Ensemble Kalman Smoother for Nonlinear Dynamics

OPTIMAL PLACEMENT AND UTILIZATION OF PHASOR MEASUREMENTS FOR STATE ESTIMATION

Transcription:

MISS DISTANCE GENERALIZED VARIANCE NON-CENTRAL CI DISTRIBUTION en Chan E-Mail: ChanAerospace@verizon.net ABSTRACT In many current practical applications, the probability o collision is oten considered as the most meaningul criterion or determining the risk to a spacecrat. owever, when the miss distance is very small (say m or less), the collision probability may not comortably serve well as the only metric or measuring the risk. In these cases, it is also important to know the statistics o the miss distance distribution, particularly the conidence limits that bound the miss distance. One approach to determine the distribution o the miss distance is to perorm rather time-consuming Monte Carlo simulations on the two conjuncting objects by choosing random initial conditions at the epoch commensurate with the covariance. The orbit propagations must be o very high precision because accuracies down to the meter range are required. The simulated results are then itted with a eibull curve which, unortunately, does not bring out the salient eatures o the distribution. The analysis here ormulates the problem analytically. For this, we consider the miss distance as given by a non-central chi distribution with unequal variances. This method eliminates the need to perorm an inordinate amount o computation, thus reducing the desired results in a timely manner by several orders o magnitude. Moreover, this generalized variance non-central chi distribution possesses the requisite eatures o the miss distance statistics. INTRODUCTION In a short-term encounter between two orbiting objects, the nominal miss distance is determined by propagating their orbits rom initial conditions at some epoch. hen they are in the vicinity o the point o closest approach, the integration steps are smaller and the minimum separation is determined using interpolation and possibly some iteration. owever, this result provides only a nominal minimum separation at the nominal time o closest approach and not a distribution o what the miss distance would be. I we desire this inormation, then we would have to perorm Monte Carlo simulations by choosing random initial conditions at the epoch through the generation o Gaussian random numbers using the covariances associated with their orbit determinations. For each such choice o initial conditions, we would have to repeat the orbit propagations and subsequent minimum separation computations. This is obviously a very timeconsuming process because we need approximately one million runs to obtain a meaningul result. An alternative approach is to propagate the two nominal orbits to some instant o time when they are both in the encounter region. In this region, the rectilinear approximation is valid just as assumed in all the collision probability computations. Thus, with this knowledge o the ingress initial conditions, we may use the analytical methods o projective geometry to determine the minimum separation and also the nominal time when this occurs. ith the ingress covariances, we can then choose seeds to perorm the Monte Carlo simulations. Note that we have eliminated the orbit propagations which orm the bulk o the previous eort. For each choice o ingress initial conditions, we use the analytical ormulation to determine the minimum separation. Note

that, even though we do not have to compute the time o occurrence as a requirement, it is not the same or all the sample runs. In this way, we obtain a distribution o miss distances. Still another approach is to ormulate the entire problem analytically. Let us irst recall the simple case o the sum o the squares o independent random variables {X i ; i =,, n}, which are all Gaussian and have zero means and also equal variances. It is well-known [] that this sum, which is a random variable, does not possess a Gaussian probability density unction (pd), but rather has a chi squared distribution. The square root o this sum has a pd which is called a chi distribution. owever, what we need to consider or our application is the case o the square root o the sum o the squares o two Gaussian random variables each with a dierent mean and a dierent variance. This is an example o the non-central chi distribution with unequal variances. Another point to make is the ollowing: I we have two Gaussian random variables with dierent means and dierent variances, then their sum or dierence is also Gaussian. This result is proved in most books on probability and statistics such as Reerence []. owever, what is not readily proved is the case o vectors when their components are Gaussian random variables. (This generalization is obvious when the covariances have principal axes which are parallel; but it is not intuitively obvious when the principal axes are not parallel.) Reerence [] takes up the diicult and laborious task o proving this statement. owever, it can also be proved in a simpler way by using the concepts o aine transormations. Finally, it may be readily proved that the covariance o the sum (or dierence) o two independent Gaussian vectors is equal to the sum o the two individual covariances. For convenience, this sum is called the Combined Covariance. (owever, in contrast, the mean o the sum is the sum o the means and the mean o the dierence is the dierence o the means.) The two results (the dierence between two independent Gaussian vectors being Gaussian, and the covariance o the dierence o two independent Gaussian vectors being the sum o the two individual covariances) are used implicitly in the present day analyses o spacecrat collision probability. Next, let us consider the nominal encounter plane deined at the instant o nominal closest approach as discussed in detail in Reerence [3]. This plane is deined to contain the relative position vector and also be orthogonal to the relative velocity vector when the two orbiting objects are at their nominal minimum separation. e assume that the position uncertainty o each object is described by a three-dimensional Gaussian pd. ence, the uncertainty o their relative position vector also has a three-dimensional Gaussian pd involving the combined covariance, being the sum o the two individual covariances. Let and denote two orthogonal coordinates in the nominal encounter plane and let the origin be at the nominal position o one (Object ) o the two orbiting objects. Let the nominal relative position vector (o Object ) be denoted by (µ, µ ). In general, the pd o the relative position uncertainty in this plane is bivariate Gaussian with an o-diagonal component. owever, or simplicity, we choose the and axes to be along the principal directions. Then, the uncertainties o the two coordinates and have Gaussian pds with standard deviations which we denote as σ and σ. Thus, the pd o the relative position uncertainty is given by an unbiased Gaussian pd as shown in Figure.

Figure Miss Distance Approach On the other hand, by invoking the concept o the non-central chi distribution, we have a pd o a random variable chi χ ( + ) deined as the square root o the sum o the squares o two Gaussian variables and. These two random variables are biased rom the origin by the means µ and µ and have standard deviations denoted by σ and σ. This is illustrated in Figure. Figure Chi Distribution Approach 3

hile the ormulation to compute the chi distribution requires some eort, this approach circumvents the need to perorm the excessive amount o computation in Monte Carlo simulations and also oers insight into the problem rom an overall perspective. NOTATION AND PRELIMINARIES In conormance with current usage [], we shall denote the Cumulative Distribution Function (CDF), also called the Probability Distribution Function (PDF), by the upper case letter F(.). Moreover, we shall denote the random variable by the upper case Y and denote a value (also called the running variable ) by the lower case y. Furthermore, we shall denote the probability that Y is less than or equal to y by P{Y y}. Thus we have the deinition Y { } F(y) PY y. () In Equation (), we also use the subscript Y in the CDF because we shall be dealing with many random variables and do not want to introduce a dierent symbol such as G or the CDF o another random variable Z. Thus, the CDF or Z would be written as F Z (z) instead o G(z). Next, we shall denote the probability density unction (pd) o a random variable Y by the lower case (.). Thus, we have the relation (y) Y df (y) dy Y =. () Let Z be a random variable with pd Z (z). Let Y be another random variable deined in terms o Z by the relation Y = Z. (3) Then, we may show that, in general, the pd Y (y) or Y is given in terms o Z (z) by [] Y(y) = Z y + Z y y. (4) Equation (4) is valid or any pd. Next, suppose Z is a normal random variable N(,), i.e., a Gaussian random variable with zero mean and unit standard variation. Then, we may show that Equation (4) becomes [] (y) = Y e y/ π y. (5) I the Gaussian random variable Z has mean µ Z and standard variation σ Z, then we may painstakingly go through two changes o variables and show that the corresponding pd is y µ Z σz e Y(y) =. (6) πσ y Z 4

It must be noted that Equation (6) is not simply obtained rom (5) by merely substituting ( y µ Z ) or y and substituting σ Z or. e must cautiously go through the process o transorming pd using the standard equation involving the Jacobian or else we obtain the incorrect answer. Except or instances when it is imperative to distinguish between the random variable Y and the running variable y (e.g., as in Equation () or the CDF), we shall break away rom the conventional notation and simply use the upper case. Another compelling reason or this simpliication is that when we use the upper case as a subscript, it is almost indistinguishable rom the lower case in regular size or some letters. Examples are (w), X (x) and Z (z). This practice o not making a distinction between variables is common in calculus in which the integration variable also takes the same notation as the upper limit in indeinite integrals. opeully, this simpliication here will not introduce unnecessary conusion in an attempt to avoid undue complications, when the notation becomes very involved. ith this change, Equation (6) takes the simpliied orm Y µ Z σz e Y(Y) = πyσ Z. (7) TE CI SQUARED CUMULATIVE DISTRIBUTION FUNCTION Consider two orthogonal coordinates and in the encounter plane deined at the nominal instant o closest approach or two objects in a short-term encounter. Since one o the objects is at the origin, thereore the miss distance is given by x = +. (8) e Suppose is Gaussian with mean µ and standard variation σ, and is also Gaussian with mean µ and standard variation σ. Furthermore, assume that and are independent random variables. This means that there is no correlation term in the bivariate pd in the (, )-plane. That is, by this choice o orientation, the and axes are in the directions o the principal axes. Thereore, and are not deined as in Reerence [3] unless the combined covariance aligns with them as the principal axes. For convenience, let us use the notation x e X (9) Z. It ollows rom Equation (7) that we have e = = X (X) ( µ ) σ πσ () 5

e = = Z(Z) ( µ ) σ πσ. () ence, Equation (8) takes the orm = X+ Z. () Because X and Z are non-negative and their sum must be equal to, it ollows that the integration with respect to (wrt) one o them, say X, must go rom zero to the value on the line X = Z. The second integration wrt Z must go rom zero to. Thus, the integration wrt to X and Z is perormed over the isosceles triangle bounded by the lines X =, Z = and X + Z =. Thereore, the CDF o is given by X+ Z w ww z X X X+ Z w w z = X Z Z Z { } F (w) = F (x+ z) = P X+ Z w = = (x) (z)dxdz (x) (z)dxdz (x)dx (z)dz. (3) The pd o is given by dierentiation o Equation (3) w w z df (w) d (w) = = X(x)dx Z(z)dz dw dw. (4) To perorm this dierentiation, we need to invoke Leibnitz s Rule (given, or example, in Reerence [4]) or dierentiation o integrals whose integrand and limits o integration involve a parameter: I(u) b(u) a(u) b(u) a(u) g(y, u)dy di(u) db(u) da(u) = g(y, u)dy g( b(u), u) g( a(u), u ). du + u du du (5) It is noted that F (w) in Equation (3) contains a double integral with two levels o upper integration limits involving the parameter w. Thus, i we apply Equation (4) to obtain the derivative o F (w), we would have to perorm two consecutive operations with Leibnitz s Rule. Even though the last term on the RS o Equation (5) vanishes because a(u) =, still there is much work involved. e shall next cast Equation (3) or computing the CDF in a slightly dierent orm. Instead o perorming the integration in terms o the variables X and Z over the isosceles triangle, we shall obtain the same result by integrating in terms o the variables and over a circle o radius centered at the origin. In this case, we use the Gaussian pds or and given by 6

( µ ) σ = e πσ (6) ( µ ) σ = e. πσ (7) Thereore, we have F () = e e d d πσ σ e recall that the error unction is deined by [3] ( µ ) σ ( µ ) σ. (8) x t er (x) = e dt π. (9) Then, Equation (8) may be written as where ( µ ) σ F () = e er Z + er Z d () πσ Z µ σ Z +µ σ. () In turn, Equation () may be written as ( µ ) σ πσ F () = e er Z + er Z d ( µ ) σ + e er ( Z) + er ( Z) d πσ ( µ ) σ ( +µ ) σ = e + e er Z + er Z d. πσ () in which, or convenience, we have used the random variable instead o the running variable w. Unlike Equation (3), it is noted that Equation () expresses the CDF F () as a single integral. e note that, by deinition, we have { } F w P w { } F w P w. (3) 7

It ollows that F w F w. (4) A word o caution is that the unctional orm on both sides o this equation are dierent because the subscripts are dierent. That is, F (.) F (.). e may prove this result more laboriously by perorming the actual substitutions into the equation or F (w) as ollows w F w w dw. (5) I we change the variable rom w to w, then the new pd w ( w) is given by dw = = = = w w w w. (6) w w w w d w ence, we obtain the desired result: w F w w dw w w = = w w d w w w = w d w F w (7) TE CI PROBABILITY DENSITY FUNCTION e shall next obtain the Chi distribution or the generalized variance non-central Chi distribution. Let us re-write the CDF in Equation () compactly as where F () = D α(, ) β(, )d. (8) D πσ α() e + e ( µ ) σ ( +µ ) σ β(,) er Z + er Z Z µ σ Z +µ σ. (9) Thereore, rom Equation (4), the pd o the Chi squared distribution is given by 8

() df () d = (3) in which, or convenience, we have used the random variable instead o the running variable w. By applying Leibnitz s Rule given by Equation (5), we obtain e note that (,) β d () = D α d+ D α β(,) +. (3) d = ( µ ) σ ( +µ ) σ µ µ α β(,) = e e er er = + + =. (3) σ σ ith some diligence, we can show that er Z i e Z = i or i = and Zi π Zi = or i = and. σ (33) It ollows that we have i= β er Zi Z = i Z i = e + e. Z Z πσ (34) ence, Equation (3) becomes ( µ ) σ ( +µ ) σ Z Z. (35) () = e + e e + e d 4πσ σ Since, thereore we can introduce a new variable θ π/ deined by Thereore, we have sin θ. (36) 9

d cosθdθ = sin θ = cos θ. (37) By substituting Equations (36) and (37) into (35), we obtain () π/ ( sinθ µ ) σ ( sinθ+µ ) σ ( cosθ µ ) σ ( cosθ+µ ) σ (38) = e + e e + e d θ. 4πσ σ This is the expression or computing the pd Chi squared distribution or the miss distance, since = x e by Equation (9). It is noted that the upper limit o integration is π/. The random variable appears only in the integrand. Even though the RS o Equation (38) involves (and not ), this is not an obstacle because is an explicit unction o. I we wish to plot the pd o the Chi squared distribution, note that the horizontal axis is and not. Next, i we wish to obtain the pd o the Chi distribution, we may invoke Equation (4) and dierentiate the RS wrt. That is, we use the ollowing analog o Equation (3) df =. (39) d owever, we can circumvent this process by using the relation (6), since we already have the pd o the Chi squared distribution given by Equation (38). Thus, we have π/ ( sinθ µ ) σ ( sinθ+µ ) σ ( cosθ µ ) σ ( cosθ+µ ) σ (4) = e + e e + e d θ. πσ σ Note that the RS o Equation (4) conveniently involves (and not ) and this greatly simpliies the display o the results. I we wish to plot the pd o the Chi distribution, note that the horizontal axis is and not. It may be shown that or the special case o zero means and equal standard deviations, we obtain the Rayleigh pd rom Equation (4). Similarly, or the case o non-zero means and equal standard deviations, we obtain the Rician pd. Appendix A gives a reormulation o Equation (38) in an alternative orm in which the expression involves and not. DISCUSSION Thus ar, we have strictly considered how to reormulate the pd o the miss distance in terms o the non-central chi distribution, but we have not justiied to how this yields the answer obtained rom the Monte Carlo simulations mentioned in the Introduction. There, we stated that when we

are in the encounter region [3], the velocities v and v o the two objects can each be treated as constant so that their motions are rectilinear and their relative motion is also rectilinear. ence, the relative velocity vector v = v v is constant in direction. Next, consider pairs o seeds or the Monte Carlo simulations chosen according to the Gaussian nature o the three-dimensional individual covariances o the two objects at ingress into the encounter region. Since the encounter plane or each pair o seeds is perpendicular to this normal vector v, it ollows that the encounter planes or all such pairs o seeds are parallel to the encounter plane or the nominal case. Moreover, because the transit time through the encounter region is very short (o the order o seconds or low orbiting objects (LEOs), the covariances o the two objects while in the encounter region can also be considered as constant. Thus, the combined covariance (being the sum o the two covariances) is constant in size, shape and orientation. For the nominal case described in the Introduction, one o the two orbiting objects is at the origin o this combined covariance or the uncertainties o the relative position vector o one object with respect to the other. That is, we have (See Figure ) + σ σ (,) = e. (4) πσ σ Let the encounter plane obtained or each pair o seeds be displaced rom the nominal one by a distance N along the direction o the normal vector v. In the (,, N) coordinate system, let the combined covariance C describing the relative position uncertainties be written as σ ρ σ σ ρ σ σ C = ρ σ σ σ ρ σ σ ρ σ σ ρ σ σ σ N N N N N N N N N. (4) From this, we obtain the inverse matrix C - and the determinant C. Then, we obtain the ollowing pd ( ρn ) ( ρn ) ( ρ )N + + σ σ σn 3(,, N) = Aexp R ( ρ ρnρn )(43) ( ρn ρρn )N ( ρn ρρn )N σσ σσn σσn where A and R are given by A = ( π) 3 Rσ σ σ N ( N N N N ) R = + ρ ρ ρ ρ ρ ρ. (44) Equation (43) gives the pd o the uncertainty o Object considering that Object is at its nominal position. Thereore, Equation (43) also gives the uncertainty o the seed or Object assuming that the seed or Object is at its relative position. Since, at the instant o closest

approach, these two seeds lie on their own encounter plane (by deinition), thereore Equation (43) also gives the uncertainty in the position o this encounter plane. Thereore, the probability dp o inding an encounter plane or the pair o seeds at a distance between N and (N + dn) rom the nominal encounter plane is given by dp = (,, N) dn. (45) 3 Consequently, i we choose Gaussian random numbers or the seeds individually according to their own covariances at ingress, the pd or the relative position vector is obtained by integrating Equation (43) with respect to N rom to +. Thus, in eect, we are obtaining the general marginal two-dimensional pd. This has already been done in Reerence [3] and the result is + ρ + ( ρ ) σ σ σ σ (,) 3(,, N)dN = e. (46) πσ σ ρ In our case, by choosing and to be along the principal directions, the correlation coeicient ρ is zero. Thereore, we obtain the pd given in Equation (4). ence, in the limit o a large number o runs, the pd o the uncertainty o the relative position vector obtained rom the Monte Carlo simulations is precisely the same as that in the nominal encounter plane as given by Equation (4). Thereore, or a given set o mean values (µ, µ ) and unequal variances (σ, σ ), the pd or the miss distance is given by the non-central chi distribution we have obtained previously. Note that not all the conjunctions o the pairs o seeds occur at the same place and at the same time. The assumptions o constant velocities and time-independent combined covariance have been crucial in obtaining the apparently obvious result above. owever, the reasoning leading to this result is rather subtle. It would have been allacious to use the ollowing argument: For each pair o seeds, we obtain an encounter plane with the pd given by Equation (4) because we have chosen the center o the combined pd to be at one o the two orbiting objects. Thereore, the pd o the relative position vector obtained rom the Monte Carlo simulations is the same as that in the nominal encounter plane as given by Equation (4). There is an error in logic in this argument: It is implicitly assumed that each seed has a covariance centered on itsel; and this is not so. In retrospect, one could ask the ollowing question: e were dealing with the three-dimensional (,, N)-space, why didn t we consider the random variable chi χ ( + + N ) deined as the square root o the sum o the squares o three Gaussian variables, and N? The reason was that we were not dealing with the pd o the unrestricted radial distance rom the origin (Generalize Figure to three dimensions.) The problem o concern was to obtain the distribution o the separation between two objects at the point o closest approach, i we perorm a large number o simulation runs using random seed inputs chosen in accordance with two threedimensional Gaussian pds, one or each seed. e established that and showed how this distribution o miss distance is obtained rom the non-central chi distribution with unequal variances involving two (not three) degrees o reedom.

NUMERICAL RESULTS Most o our attention has been directed to obtaining Equation (4) or the pd o the random variable χ ( + ), because that was what we sought or the distribution o the miss distance. An example o a pd Chi distribution is shown in Figure 3 or the sample case o σ =.5 km, σ =. km, µ =.5 km, µ =. km. Figure 3 Example o Probability Density Function o Miss Distance It turns out, however, that the more useul representation or our application is actually the Cumulative Distribution Function o the Chi random variable χ ( + ), because it gives the probability the miss distance does not exceed some value. Furthermore, the inormation contained in this CDF o χ is exactly the same as that contained in the CDF o the Chi Squared random variable χ +, because the latter gives the probability the miss distance squared does not exceed the square o that value. Figure 4 shows the CDF o χ or the same sample case as that in Figure 3. An example application is the determination o the probability or which the miss distance does not exceed a certain speciied value, say α. For this, we need to ind the area under the curve in Figure 3 rom to α. owever, we can obtain the same inormation by merely obtaining the probability corresponding to α in the curve in Figure 4. Using this approach, we can ind the lower and upper values o the miss distance, excluding the probability below.5 and the probability above.975, so as to give us a 95% conidence interval. Figure 5 depicts these limits or the nominal miss distance obtained rom Two Line Element sets or the ive days preceding the collision o Iridium 33 and osmos 5 on February, 9. 3

Figure 4 Example o Cumulative Distribution Function o Miss Distance Squared -3 Iridium-osmos Collision Close Approach Statistics -4 Probability -5-6 -7 8 6 4 Miss Distance (km) -8-9 -5.5-5 -4.5-4 -3.5-3 -.5 - -.5 - -.5 Elset (days beore conjunction) Figure 5 Miss Distance and Collision Probability or the Iridium 33/osmos 5 Collision 4

Finally, it remains to veriy that the miss distance distributions obtained analytically agree with those obtained rom the Monte Carlo simulations. A number o plots comparing these results showing good agreement have been obtained. Figure 6 illustrates one particular case o this comparison. Figure 6 Comparison o Statistical Miss Distance Distributions CONCLUSIONS () The analytical distributions agree very well with those obtained rom Monte Carlo simulations. By the use o the pd Chi distribution or the CDF Chi Squared distribution, we have circumvented the direct approach utilizing time-consuming high precision orbit propagation to obtain miss distances when they are very small (o less than m). () The Chi distribution applicable to our analysis is the one involving two degrees o reedom, not that involving three degrees o reedom, even though we have three-dimensional covariances associated with each o the two orbiting objects. REFERENCES [] Sheldon Ross, A First Course in Probability, Prentice all, p 3,, 55, 8 th ed (). [] T.. Anderson, An Introduction to Multivariate Statistical Analysis, John iley & Sons, pp 3-33 (3). [3] F. enneth Chan, Spacecrat Collision Probability, The Aerospace /AIAA Presses, pp 7-37, 47-6, 39 (8). [4] Philip Franklin, Methods o Advanced Calculus, McGraw-ill, pp 5-5 (944). 5

APPENDIX A: ALTERNATIVE FORM OF EQUATION (38) By multiplying out the squares in the exponents in the integrand in Equation (38), we may write this equation as () = 4 πσ σ π/ µ ( σ +µ σ ) e ( sin θ σ ) + cos θ σ µ sin θ σ µ sin θ σ µ cosθ σ µ cosθ σ e e e e e + + dθ. (A) µ ( σ +µ σ ) = e πσ σ π/ ( sin θ σ + cos θ σ ) µ sin θ µ cosθ e cosh cosh σ d θ σ For convenience, let us introduce the ollowing notation µ sin θ µ cosθ x = and y =. (A) σ σ Then, we have cosh x cosh y e e e e 4 x y x y x y x y ( e + e e + e = + + + ) 4 = cosh ( x + y ) + cosh ( x y ). x x y y = ( + )( + ) (A3) Thereore, Equation (A) may be written as () = πσ σ π/ µ ( σ +µ σ ) e ( sin θ σ + cos θ σ ) e cosh( x+ y) + cosh( x y) d θ. (A4) In view o the expansions 4 6 (x + y) (x + y) (x + y) cosh( x + y) = + + + +! 4! 6! 4 6 (x y) (x y) (x y) cosh( x y) = + + + +,! 4! 6! (A5) 6

we obtain ( + ) + ( ) cosh x y cosh x y 4 4 6 4 4 6 (x + y ) (x + 6x y + y ) (x + 5x y + 5x y + y ) = + + + +.! 4! 6! (A6) e note that the RS o Equation (A6) involves x and y to the even powers only. This means that occurs in integral powers only in the integrand o Equation (A4). That is, there are actually no terms with hal powers in in the pd o the Chi squared distribution () Equation (38). Equation (A4) oers an alternative way to compute the pd o the Chi squared distribution. 7