Numerical Regularization for Atmospheric Inverse Problems

Similar documents
Superlinear Parabolic Problems

Schrödinger Equations and Diffusion Theory

Observation and Control for Operator Semigroups

Quantum Mechanics of Molecular Structures

Solving Algebraic Computational Problems in Geodesy and Geoinformatics

Controlled Diffusion Processes

de Gruyter Textbook Distributions Generalized Functions with Applications in Sobolev Spaces Bearbeitet von Pulin Kumar Bhattacharyya

Seismic Wave Propagation and Scattering in the Heterogeneous Earth : Second Edition

The Mathematics of Minkowski Space-Time

Analysis II. Bearbeitet von Herbert Amann, Joachim Escher

Supersymmetry and Equivariant de Rham Theory

Size Exclusion Chromatography

Feynman-Kac-Type Theorems and Gibbs Measures on Path Space

International Steam Tables - Properties of Water and Steam based on

Reflecting Telescope Optics II

Perspectives on Projective Geometry

Differential Scanning Calorimetry

An Introduction to the Theory of Functional Equations and Inequalities

Introduction to Wave Scattering, Localization and Mesoscopic Phenomena

The Quantum Mechanics Solver

An Introduction to Kinetic Monte Carlo Simulations of Surface Reactions

Waves Called Solitons

Full-Potential Electronic Structure Method

Geometry by Its History

Ultrafast Spectroscopy of Semiconductors and Semiconductor Nanostructures

The Beauty of Everyday Mathematics

Symmetries and Group Theory in Particle Physics

Geological Methods in Mineral Exploration and Mining

Beam-Wave Interaction in Periodic and Quasi-Periodic Structures

Computational Chemistry Workbook

Fiber Optics. Physics and Technology. Bearbeitet von Fedor Mitschke

Scanning Probe Microscopy

Arc-Continent Collision

The Ecozones of the World

Quantum Kinetics in Transport and Optics of Semiconductors

Stochastic Variational Approach to Quantum-Mechanical Few-Body Problems

Laser Processing of Materials

Design and Construction of Tunnels

Satellite Orbits. Models, Methods and Applications. Bearbeitet von Oliver Montenbruck, Eberhard Gill

Space-Time Reference Systems

Turbomachinery Flow Physics and Dynamic Performance

Computational Acoustics of Noise Propagation in Fluids - Finite and Boundary Element Methods

The Geology of Stratigraphic Sequences

Study and Design of Differential Microphone Arrays

Gas Cyclones and Swirl Tubes

Piezoelectric Multilayer Beam Bending Actuators

Differential Geometry of Lightlike Submanifolds

The BCS-BEC Crossover and the Unitary Fermi Gas

Dynamics of Fibre Formation and Processing

The k p Method. Electronic Properties of Semiconductors. Bearbeitet von Lok C Lew Yan Voon, Morten Willatzen

Mathematical Methods for Mechanics

Impacts of Global Change on the Hydrological Cycle in West and Northwest Africa

Mass Extinction. Bearbeitet von Ashraf M.T Elewa

Ray Optics, Fermat s Principle, and Applications to General Relativity

Apartness and Uniformity

Three-Dimensional Free-Radical Polymerization

Interfacial Electrochemistry

Dynamical Systems, Ergodic Theory and Applications

Groundwater in Ethiopia

Modern Developments in X-Ray and Neutron Optics

Optimal Time-Domain Noise Reduction Filters

Adsorbed Species on Surfaces and Adsorbate-Induced Surface Core Level Shifts

Turbomachinery Flow Physics and Dynamic Performance

Exploration Geophysics

The Middle Paraná River

Plant Cell Monographs 13. The Chloroplast. Interactions with the Environment. Bearbeitet von Anna Stina Sandelius, Henrik Aronsson

Thermal Stress Resistance of Materials

A Comprehensive Treatment of q-calculus

Intraseasonal Variability in the Atmosphere-Ocean Climate System

Structured Population Models in Biology and Epidemiology

Flow and Transport in Porous Media and Fractured Rock

Tropical Rainforest Responses to Climatic Change

Springer Monographs in Mathematics. Algebraic Cobordism. Bearbeitet von Marc Levine, Fabien Morel

Plane Algebraic Curves

Physics, Formation and Evolution of Rotating Stars

Quartz: Deposits, Mineralogy and Analytics

Organic Semiconductors in Sensor Applications

Groundwater Geophysics

Particle Accelerators, Colliders, and the Story of High Energy Physics

Matrix Information Geometry

Mycorrhiza. State of the Art, Genetics and Molecular Biology, Eco-Function, Biotechnology, Eco-Physiology, Structure and Systematics

Natural Time Analysis: The New View of Time

DS-GA 1002 Lecture notes 10 November 23, Linear models

Experimental and Computational Solutions of Hydraulic Problems

LEP - The Lord of the Collider Rings at CERN

Quantum Field Theory

Quantum Thermodynamics

Symplectic Invariants and Hamiltonian Dynamics

Solving Differential Equations in R

Transport in Metal-Oxide-Semiconductor Structures

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

Principles of Sedimentary Basin Analysis

Satellite Altimetry for Geodesy, Geophysics and Oceanography

Particle Accelerators, Colliders, and the Story of High Energy Physics

Least Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo

Introduction to Modern Number Theory

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

The Calculus of Computation

Elementary linear algebra

Transcription:

Springer Praxis Books Numerical Regularization for Atmospheric Inverse Problems Bearbeitet von Adrian Doicu, homas rautmann, Franz Schreier st Edition. 2. Buch. xiii, 426 S. Hardcover ISBN 978 3 642 5438 9 Format B x L): 6,8 x 24 cm Gewicht: 95 g Weitere Fachgebiete > Mathematik > Numerik und Wissenschaftliches Rechnen > Angewandte Mathematik, Mathematische Modelle Zu Inhaltsverzeichnis schnell und portofrei erhältlich bei Die Online-Fachbuchhandlung beck-shop.de ist spezialisiert auf Fachbücher, insbesondere Recht, Steuern und Wirtschaft. Im Sortiment finden Sie alle Medien Bücher, Zeitschriften, CDs, ebooks, etc.) aller Verlage. Ergänzt wird das Programm durch Services wie Neuerscheinungsdienst oder Zusammenstellungen von Büchern zu Sonderpreisen. Der Shop führt mehr als 8 Millionen Produkte.

2 Ill-posedness of linear problems Inverse problems typically involve the estimation of certain quantities based on indirect measurements of these quantities. he inversion process is often ill-posed in the sense that noise in the data gives rise to significant errors in the estimate. In this chapter we introduce the concept of ill-posedness and analyze the solvability and ill-posedness of linear discrete equations. Our analysis is focused on a classical example in atmospheric remote sensing, namely the temperature retrieval by nadir sounding. In a continuous setting, this retrieval problem is modeled by a Fredholm integral equation of the first kind, which is the prototype of ill-posed problems. 2. An illustrative example o explain the difficulties associated with the solution of linear inverse problems, we consider measuring a temperature profile by nadir sounding. Spaceborne remote sensing of atmospheric temperature uses absorption features of gases with well-known and constant mixing ratios as for instance, the CO 2 bands at 5 and 4.3 μm in the infrared or the O 2 lines at 6 GHz in the microwave regime. In a plane-parallel atmosphere and under the assumption that the background contribution from the surface can be neglected, the diffuse radiance at the top of the atmosphere z = z max and in a direction with zero scan angle is given by the Schwarzschild equation I ν) = zmax B ν, z)) z ν, z) dz. In the microwave region, the Rayleigh Jeans approximation allows the following representation of the Planck function B ν, z))=2ck B ν 2 z). As a result and when neglecting the temperature dependence of the transmission, we obtain I ν) = zmax k ν, z) z) dz, 2.)

24 Ill-posedness of linear problems Chap. 2 with k ν, z) =2ck B ν 2 ν, z) 2.2) z being the kernel function. Equation 2.), which we rewrite in the generic form y = I and x = ) y ν) = zmax k ν, z) x z) dz, ν [ν min,ν max ], 2.3) is a Fredholm integral equation of the first kind and represents the mathematical model of a continuous problem. o formulate our problem in a general setting, we consider the space of real-valued, square integrable functions on the interval [a, b], denoted by L 2 [a, b]). Actually, L 2 [a, b]) is a Hilbert space under the inner product u, v = b u t) v t) dt a b and the induced norm u = a ut)2 dt. Assuming that y belongs to the Hilbert space Y = L 2 [ν min,ν max ]) and x belongs to the Hilbert space X = L 2 [,z max ]), we introduce the linear operator K : X Y by the relation Kx = zmax he integral equation 2.3) can then be written as k,z) x z) dz. Kx = y, 2.4) and since cf. 2.2)) k L 2 [ν min,ν max ] [,z max ]), the linear operator K is bounded and compact the image of any bounded sequence of functions in L 2 has at least one converging subsequence). A spectral instrument cannot measure a continuous signal and the data is a collection of discrete observations. More specifically, the radiances y ν i )=I ν i ), with {ν i },m being an equidistant set of points in the spectral interval [ν min,ν max ], represent the data, and the equation to be solved takes the form y ν i )= zmax k ν i,z) x z) dz, i =,...,m. 2.5) he semi-discrete equation 2.5) is a mathematical model for discrete observations of a physical process and can be expressed in compact form as K m x = y m. 2.6) he data y m is a vector with entries [y m ] i = y ν i ), i =,...,m, and K m is a linear operator acting between the Hilbert space X and the finite-dimensional Euclidean space R m, zmax [K m x] i =Kx)ν i )= k ν i,z) x z) dz.

Sect. 2. An illustrative example 25 he discretization approach which transforms the continuous equation 2.3) into the semidiscrete equation 2.5) is known as the collocation method. It should be pointed out that the collocation method can be regarded as a projection method with delta functions as basis functions. For a complete discretization, we consider the subspace X n X with the not necessarily orthonormal) basis {Φ nj } j=,n and define the approximation or the interpolant x n X n of x as the solution of the equation Representing x n as a linear combination of basis functions, x n = ξ j Φ nj, we obtain the system of equations y ν i )= j= [ zmax In matrix form, 2.8) can be written as K m x n = y m. 2.7) j= ] k ν i,z)φ nj z) dz ξ j, i =,...,m. 2.8) K mn x n = y m, 2.9) where x n =[ξ,...,ξ n ] is the coordinate vector and the matrix K mn, with entries [K mn ] ij =[K m Φ nj ] i =KΦ nj )ν i )= zmax k ν i,z)φ nj z) dz, is a linear map between the finite-dimensional Euclidean spaces R n and R m. he discretization approach which transforms the continuous equation 2.3) into the discrete equation 2.8) is called a projection method. Let us make some comments on the choice of the set of basis functions {Φ nj } for representing the state parameter x. ) If {z j } j=,n is a discretization grid of the altitude interval [,z max ] with z =and z n = z max, we may choose the piecewise constant functions P nj z) = {, zj z<z j,, otherwise, j =,...,n as basis functions. Using the orthogonality relations P ni,p nj =, i j, and P nj 2 = z j z j, we obtain ξ j = z j z j x n,p nj = z j z j zj z j x n z) dz for j =,...,n. hus, the entries of the coordinate vector are the mean values of the atmospheric profile over each altitude interval layer).

26 Ill-posedness of linear problems Chap. 2 2) For the discretization grid {z j } j=,n+ with z = z =and z n = z n+ = z max, the piecewise linear functions or hat functions), z z j ) / z j z j ), z j <z z j, H nj z) = z j+ z) / z j+ z j ), z j z<z j+, j =,...,n,, otherwise, can also be chosen as basis functions. Since H nj z i )= {, i = j,, i j, i, j =,...,n, it follows that ξ j = x n z j ) for j =,...,n, and we conclude that the entries of the coordinate vector are the values of the atmospheric profile at each grid point level). 3) For a smoother and more accurate approximation, we have to use a piecewise polynomial approximation with higher-order pieces than broken lines. he most popular choice is the B-spline interpolation de Boor, 2). In this case, for the discretization grid {z j } j=,n with z =and z n = z max, x n is expressed as x n z) = ξ j B nkj z), j= where B nkj are the B-splines of order k. Note that B nkj is a piecewise polynomial of degree of at most k, and that the B nkj, j =,..., n, are locally linear independent. A well-conditioned basis of B-splines can be obtained with the recursion formulas {, zj z<z B nj z) = j+,, otherwise, z t j B nkj z) = B nk j z)+ t j+k z B nk j+ z), k 2, t j+k t j t j+k t j+ where t t 2... t n+k are the knots at which the polynomials are tied together by the continuity conditions. In many problems, where extrapolation beyond z =and z = z max is not anticipated, it is a common practice to set and t = t 2 =...= t k = t n+ = t n+2 =...= t n+k = z max. he second-order B-spline B n2j coincides with the hat functions, and for this reason, B n2j is also called a linear B-spline.

Sect. 2.3 Concept of ill-posedness 27 2.2 Concept of ill-posedness he mathematical formulation of inverse problems leads to equations that typically are ill-posed. According to Hadamard, the equation Kx = y, 2.) with K being a linear operator acting from the Hilbert space X into the Hilbert space Y, is called well-posed provided Engl et al., 2; Rieder, 23; Vogel, 22) ) for any y Y, a solution x exists; 2) the solution x is unique; 3) the solution is stable with respect to perturbations in y, in the sense that if Kx = y and Kx = y, then x x whenever y y. Equivalently, equation 2.) is called well-posed if the operator K is bijective and the inverse operator K is continuous. As equation 2.) is a mathematical model of a continuous problem, the term well-posed is also used when referring to the underlying problem. If one of Hadamard s conditions is violated, the problem is called ill-posed. Denoting by R K) ={Kx/x X} the range space of K and by N K) ={x X/Kx =} the null space of K, it is apparent that Kress, 999) ) if K is not surjective R K) Y ), then equation 2.) is not solvable for all y Y non-existence); 2) if K is not injective N K) ), then equation 2.) may have more than one solution non-uniqueness); 3) if K exists but is not continuous, then the solution x of equation 2.) does not depend continuously on the data y instability). Non-existence can occur in practice because the forward model is approximate or because the data contains noise. Non-uniqueness is introduced by the need for discretization and is a peculiarity of the so-called rank deficient problems, characterized by a matrix K mn with a non-trivial null space. In particular, state vectors x that lie in the null space of K mn solve the equation K mn x =, and by superposition, any linear combination of these null-space solutions can be added to a particular solution and does not change the fit to the data. Violation of the third Hadamard condition creates serious numerical problems because small errors in the data space can be dramatically amplified in the state space. When a continuous ill-posed problem is discretized, the underlying discrete problem inherits this ill-posedness and we say that we are dealing with a discrete ill-posed problem Hanke and Hansen, 993). he ill-posedness of a discrete linear problem, written in the form of a linear system of equations, is reflected by a huge condition number of the coefficient matrix. In this regard, the term ill-conditioned system of equations is also used to describe instability. o stabilize the inversion process we may impose additional constraints that bias the solution, a process that is generally referred to as regularization.

28 Ill-posedness of linear problems Chap. 2 2.3 Analysis of linear discrete equations he Fredholm integral equation Kx = y is severely ill-posed, when K is a compact operator with an infinite-dimensional range space Engl et al., 2). his means that the inverse operator K is unbounded and that the third Hadamard condition is not fulfilled. An analysis of continuous ill-posed problems regarding their solvability and ill-posedness is given in Appendix A; here we pay attention to the discrete case. From a strictly mathematical point of view, the discrete equation 2.9), written in the more familiar form as Kx = y, 2.) is well-posed, as any nonsingular matrix automatically has a continuous inverse. However, in terms of condition numbers, the fact that a continuous problem is ill-posed means that the condition number of its finite-dimensional approximation grows with the quality of the approximation Hanke and Hansen, 993). Increasing the degree of discretization, i.e., increasing the approximation accuracy of the operator, will cause a huge condition number of the matrix and a dramatic amplification of rounding errors. As a result, the approximate solution becomes less and less reliable. 2.3. Singular value decomposition In order to demonstrate the ill-posed nature of the discrete equation 2.) we first introduce the concept of singular value decomposition of a matrix. For an m n matrix K, the matrix K K is symmetric and positive semidefinite, and as a result, the eigenvalues of K K are real and non-negative. he non-negative square roots of these eigenvalues are called the singular values of K. If rank K) =r, the matrix K has exactly r positive singular values counted according to their geometric multiplicity. o simplify our exposition we suppose that these singular values are simple and throughout this book, the claim rank A) =r tacitly assumes that the matrix A has exactly r positive and distinct singular values. Note that a symmetric matrix A is said to be positive definite if x Ax > for all x, and positive semidefinite if x Ax. All eigenvalues of a symmetric and positive definite matrix are positive real numbers. Also note that the rank of a matrix A is the maximal number of linearly independent column vectors of A column rank), or the maximal number of linearly independent row vectors of A row rank). If K is of rank r, and {σ i },n denotes the set of singular values appearing in decreasing order, σ >σ 2 >...>σ r >σ r+ =...= σ n =, then there exist the orthonormal sets {v i },n R n and {u i },m R m such that and Kv i = σ i u i, K u i = σ i v i, i =,...,r, 2.2) Kv i =, i = r +,...,n, K u i =, i = r +,...,m. 2.3) Each system σ i ; v i, u i ) with these properties is called a singular system of K. he sets {v i },r and {v i } i=r+,n are orthonormal bases of N K) and N K), respectively,

Sect. 2.3 Analysis of linear discrete equations 29 i.e., N K) = span {v i },r, N K) = span {v i } i=r+,n, 2.4) while {u i },r and {u i } i=r+,m are orthonormal bases of R K) and R K), respectively, i.e., R K) = span {u i },r, R K) = span {u i } i=r+,m. 2.5) In 2.4) and 2.5), N K) and R K) are the orthogonal complements of N K) and R K), respectively, and we have the representations R n = N K) NK) and R m = R K) RK), where the notation stands for the direct sum of two sets, A B = {x + y/ x A, y B}. he condition number of the matrix K is defined as the ratio of the largest to the smallest singular value, that is, κ K) =σ /σ r. Equations 2.2) 2.3) can be written in matrix form as K = UΣV, 2.6) where U =[u,...,u m ] and V =[v,...,v n ] are orthogonal or orthonormal) m m and n n matrices, respectively, and Σ is an m n matrix of the form [ ] diag σi ) Σ= r r, with diag σ i ) r r being an r r diagonal matrix. he representation 2.6) is called the singular value decomposition SVD) of the matrix K. A positive definite matrix A is nonsingular and its singular value decomposition coincides with its spectral decomposition, that is, A = VΣV. hroughout this book the discussion of positive definite matrices is restricted to symmetric matrices, although a general real matrix is positive definite if and only if its symmetric part /2) A + A ) is positive definite, or equivalently, if and only if its symmetric part has all positive eigenvalues. Hence, when we say that A is positive definite, in fact, we mean that A is symmetric and positive definite. Positive definite matrices are important in statistics essentially because the covariance matrix of a random vector is always positive definite, and conversely, any positive definite matrix is the covariance matrix of some random vector in fact, of infinitely many). For A = VΣV, the square root of A is taken as A /2 = VΣ /2 V, and evidently A /2 is symmetric. However, a positive definite matrix has many non-symmetric square roots, among which the one obtained by the Cholesky factorization A = L L, where L is upper triangular, is of particular interest. 2.3.2 Solvability and ill-posedness Let K be an m n matrix of rank n with the singular system {σ i ; v i, u i )}. he assumption rank K) =nyields N K) =, which, in turn, implies that the linear operator K is injective. he solvability of equation 2.) is stated by the following result: the linear equation 2.) is solvable if and only if y RK), and the unique solution is given by x = u σ i y ) v i. 2.7) i

3 Ill-posedness of linear problems Chap. 2 o prove this result we first assume that x is a solution of 2.), i.e., Kx y N K ), we see that y y = x K y =, = y. If and since R K) =N K ), the necessity of condition y RK) follows. Conversely, let y RK). hen, y can be expressed in terms of the orthonormal basis {u i },n of R K) as follows: y = u i y ) u i. For x defined by 2.7), we have cf. 2.2)) Kx = σ i u i y ) Kv i = u i y ) u i, and we deduce that Kx = y. Finally, the uniqueness of x follows from the injectivity of K. Equation 2.7) defines a linear operator K : R m R n by the relation K y = which also allows a representation by an n m matrix, K = σ i u i y ) v i, y R m, 2.8) σ i v i u i. his operator or matrix, which maps y RK) into the solution x of equation 2.), that is, x = K y, is called the generalized inverse. By convention, the data vector y which belongs to the range space of K, will be referred to as the exact data vector, and x = K y will be called the exact solution. In practice, the exact data is not known precisely and only the noisy data is available. he noisy data vector y δ is a perturbation of the exact data vector y, and we have the representation y δ = y + δ, where δ is the instrumental noise. In general, y δ R m, and there is no guarantee that y δ RK). As a result, the linear equation is not solvable for arbitrary noisy data and we need another concept of solution, namely the least squares solution. For the noisy data vector m y δ = u i y δ) u i, 2.9)

Sect. 2.3 Analysis of linear discrete equations 3 the least squares solution of the linear equation 2.) is defined by x δ = σ i u i y δ) v i. 2.2) he least squares solution can be characterized as follows: ) the image of x δ under K is the projection of y δ onto R K), that is, Kx δ = P RK) y δ = u i y δ) u i ; 2) x δ has the optimality property 3) x δ solves the normal equation x δ = arg min x y δ Kx ; K Kx = K y δ. Assertion ) follows from 2.2) in conjunction with 2.2). Considering 2), we see that y δ Kx δ = y δ P RK) y δ = min y δ y = min y δ Kx x y RK) and the conclusion is apparent. For proving 3), we use 2.9) and 2.2) to obtain y δ Kx δ = m i=n+ u i y δ) u i. hus, y δ Kx δ RK) = N K ) ; this gives K y δ Kx δ) =and the proof is complete. By virtue of 2.8) and 2.2), the least squares solution can be expressed as x δ = K y δ, and since x δ solves the normal equation, the SVD of K yields the factorization K = K K ) K = VΣ U, 2.2) with [ Σ = diag σ i )n n ]. Note that for rank K) =r<n, x δ defined by x δ = r σ i u i y δ) v i,

32 Ill-posedness of linear problems Chap. 2 is an element of N K) = span {v i },r and represents the unique least squares solution of equation 2.) in N K).Ifx is an arbitrary vector in N K), then K x δ + x ) = PRK) y δ, and x δ + x is a least squares solution of equation 2.) in R n. Using the orthogonality relation x x δ =, we observe from x δ + x 2 = x δ 2 +2x x δ + x 2 = x δ 2 + x 2, that x δ represents the least squares minimal norm solution of equation 2.). For discrete ill-posed problems, the following features of the singular values and vectors are relevant Hansen, 998): ) as the dimension of K increases, the number of small singular values also increases; 2) as the singular values σ i decrease, the corresponding singular vectors u i and v i have more sign changes in their components. As a consequence of the oscillatory behavior of the high-order singular vectors, the norm of the least squares solution becomes extremely large and x δ is not a reliable approximation of x. o be more concrete, we choose a large singular-value index i and consider a perturbation of the exact data vector y in the direction of the singular vector u i, y δ = y +Δu i, with Δ= y δ y being the noise level. he least squares solution is then given by and the ratio x δ = x + Δ σ i v i x δ x y δ y = σ i is very large if σ i is very small Figure 2.). In this context, any naive approach which tries to compute x δ by using 2.2) will usually return a useless result with extremely large norm. he instability of an ill-conditioned linear system of equations depends on the decay rate of the singular values. In this sense, we say that a discrete problem is mildly ill-posed if σ i = Oi β ) for some positive β, and severely ill-posed if σ i = Oe i ). 2.3.3 Numerical example he difficulties associated with the solution of ill-posed problems will be demonstrated by considering the temperature nadir sounding problem 2.). As water absorption is dominant for frequencies below 4 GHz and above 2 GHz, we assume a single oxygen line at position ˆν O2 and ignore other absorbers completely.

Sect. 2.3 Analysis of linear discrete equations 33 Fig. 2.. he generalized inverse K maps the exact data vector y into the exact solution x and the noisy data vector y δ into the least squares solution x δ. he image of x δ under K is the projection of y δ onto R K). Although the error in the data space Δ= y δ y is small, the error in the state space x δ x can be very large. Neglecting the temperature-dependence of line strength and pressure broadening, and assuming an observer at infinity gives the absorption optical depth omitting the gas index) τ abs ν, z) = π z Sγ L n z) p z) [ ] 2 dz. 2.22) ν ˆν) 2 p z) p ref + γ L p ref he volume mixing ratio q of O 2 is constant with altitude, i.e., q =.2, and the number density depends on altitude through pressure and temperature, n z) =q p z) k B z). 2.23) aking into account that pressure varies approximately exponentially with altitude see.)), assuming p ref = p and ignoring the altitude dependence of the temperature in 2.23) varies between 2 and 3 K), the integral in 2.22) can be evaluated analytically. he result is ν ˆν) 2 + γl 2 exp 2z H ) τ abs ν, z) =a log ν ˆν) 2, and the transmission is then given by a ν, z) =exp τ abs ν, z)) = ν ˆν) 2 ν ˆν) 2 + γl 2 exp 2z H ), with a = qsp H 2πk B γ L = qsp 2πγ L mg.

34 Ill-posedness of linear problems Chap. 2 8 6 5 Altitude [km] 6 4 4 3 2 2.2.4.6.8 ransmission.2.4.6.8 d ransmission / dz Fig. 2.2. ransmission left) and weighting function / z right) for a temperature nadir sounding model with exponent a =., line position ˆν = 2. cm and δν = ν ˆν = 6,..., 2 cm. Chosing S =.5 25 cm / molec cm 2), γ L =. cm, m =5 23 g, and p = 6 gcm s 2, we find that a. he transmission and the weighting function, defined by / z, are depicted in Figure 2.2. Now, passing from the vertical coordinate z to the non-dimensional coordinate ζ = 2z H and making the change of variable 2ck B γl 2 I ν) I ν), the integral equation 2.) becomes with I ν) = ζmax K ν, ζ) = K ν, ζ) ζ) dζ, ν [ν min,ν max ], ν 2 ν ˆν) 2 exp ζ) [ν ˆν) 2 + γ 2 L exp ζ) ] 2.

Sect. 2.3 Analysis of linear discrete equations 35 Assuming a discretization scheme with piecewise constant functions we obtain the discrete equation [I m ] i = n ζ) = ) ζ j Pnj ζ) j= [K mn ] ij [ n ] j, i =,...,m, 2.24) j= with the forward model matrix N is the number of quadrature points) [K mn ] ij = the data vector and the state vector ζmax K ν i,ζ) P nj ζ) dζ = N K ) ) ν i, ζ k Pnj ζk ΔζN, 2.25) k= [I m ] i = I ν i ), [ n ] j = ζ j ). he layer centerpoints in the data and state spaces are ν i = i ) Δν m, i =,...,m, 2 and ζ j = j ) Δζ n, j =,...,n, 2 respectively, while the discretization steps are Δν m = ν max ν min ) /m and Δζ n = ζ max /n. he integration scheme for computing the integral in 2.25) does not depend on the discretization grid in the state space and we have ζ k = k ) Δζ N, k =,...,N, 2 with Δζ N = ζ max /N. Further, we set ν min =.98 cm, ν max =2.2 cm, ζ max =5 and choose the exact temperature profile as 22+28 5 2 ζ), ζ 2.5, 22, 2.5 <ζ 5, ζ) = 22 + 25 3 ζ 5), 5 <ζ, 2.26) 27, <ζ 4, 25 + 2 5 ζ), 4 <ζ 5.

36 Ill-posedness of linear problems Chap. 2 Our analysis is organized as follows: ) we fix the number of discrete data and quadrature points, m = 2 and N =, respectively, and compute the exact data vector by using the relation [I m ] i = ζmax K ν i,ζ) ζ) dζ = 2) we generate the noisy data vector I δ m as N K ) ν i, ζ k ) ζ k ΔζN ; k= I δ m = I m + δi, where δi is a random Gaussian vector with zero mean and covariance C δ = σ 2 I m ; the noise standard deviation is defined in terms of the signal-to-noise ratio SNR by σ = I m msnr ; 3) for different values of the discretization index n, we compute the least squares solution and determine the solution error where ε 2 n = n δ 2 = δ n ζ) = δ n = K mni δ m, ζmax [ ] δ n j= [ ζ) δ n ζ) ] 2 dζ, j P nj ζ). In the left panel of Figure 2.3 we plot the condition number of the matrix K mn for different values of the discretization index n. As expected, increasing the degree of discretization causes a huge condition number of the matrix and a dramatic amplification of solution errors is expected. he behavior of the Picard coefficients Pi δ = u i Iδ m, σ i which reflects the ill-posedness of the discrete equation, is shown in the right panel of Figure 2.3. As we will see in Chapter 3, if the sequence of Picard coefficients does not decay on average, then the reconstruction error is expected to be extremely large. For n = 2, the discrete Picard condition is satisfied on average and we anticipate reasonable small errors, while for n =8, the Picard coefficients do not decay on average and extremely large reconstruction errors are expected. Figure 2.4 shows the singular vectors v, v 6 and v 4 corresponding to the singular values σ =.8 3, σ 6 =.2 2 and σ 4 =.6, respectively. he results illustrate that the number of oscillations of the singular vector components increases when the corresponding singular values decrease. Note that fine structures in the profile are reproduced by singular vectors corresponding to smaller singular values, while singular vectors corresponding to larger singular values are responsible for the rough structures see, e.g., Rodgers, 2).

Sect. 2.3 Analysis of linear discrete equations 37 5 n=2 n=8 8 4 Condition Number 6 Picard Coefficients 3 4 2 2 2 4 6 8 2 Discretization Index Fig. 2.3. Left: condition number versus the discretization index. n =2and n =8, in the case SNR =. 5 5 5 2 Singular Value Index Right: Picard coefficients for Altitude ζ 5.5.5 v.5.5 v 6.5.5 v 4 Fig. 2.4. Singular vectors v, v 6 and v 4 for n =8. In the left panel of Figure 2.5 we plot the solution errors versus the discretization index n for different values of the SNR. he results show that the solution error has a minimum for an optimal value n of the discretization index. hese values are n =2 for SNR =, n =3for SNR = 5, and n =4for SNR = ; thus n increases

38 Ill-posedness of linear problems Chap. 2 5 SNR = SNR = 5 SNR = 5 Relative Error Altitude ζ 5 retrieved n=2) retrieved n=4) exact 5 2 4 6 8 2 Discretization Index 2 3 4 5 emperature [K] Fig. 2.5. Left: relative errors versus the discretization index. Right: retrieved profiles for n =2 and n =4, in the case SNR =. when the SNR increases. he solution errors corresponding to the optimal values of the discretization index are reasonable small. his is apparent from the right panel of Figure 2.5, which illustrates the retrieved profile for n =2and SNR =. In contrast, up to the discretization index n =4, the least squares solution oscillates and has a large norm. he behavior of the solution error illustrates that projection methods have an inherent regularizing property Natterer, 977). If the discretization is too coarse, the finitedimensional equation will be fairly well conditioned, but the solution will be affected by a large discretization error ε 2 dn = n 2, where n ζ) = [ ] n j= j P nj ζ) and n = K mni m is the least squares solution in the noise-free case. On the other hand, if the discretization is too fine, then the influence of the small singular values is significant, and the so-called noise error ε 2 nn = n n δ 2 explodes. Both the discretization and the noise errors contribute to the solution error ε n, and the optimal discretization index n balances the two error components. he main drawback of regularization by projection is that the optimal discretization index n is too small and the corresponding vertical resolution is too low. Regularization methods yielding satisfactory reconstruction errors with high vertical resolutions are the topic of the next chapters.