Unbiased Risk Estimation as Parameter Choice Rule for Filter-based Regularization Methods

Similar documents
Empirical Risk Minimization as Parameter Choice Rule for General Linear Regularization Methods

Statistical Inverse Problems and Instrumental Variables

Inverse problems in statistics

Convergence rates of spectral methods for statistical inverse learning problems

Statistical regularization theory for Inverse Problems with Poisson data

Iterative regularization of nonlinear ill-posed problems in Banach space

Inverse problems in statistics

ORACLE INEQUALITY FOR A STATISTICAL RAUS GFRERER TYPE RULE

Regularization in Banach Space

MURDOCH RESEARCH REPOSITORY

Numerical differentiation by means of Legendre polynomials in the presence of square summable noise

Inverse problems in statistics

Iterative Regularization Methods for Inverse Problems: Lecture 3

Regularization and Inverse Problems

CONVERGENCE RATES OF GENERAL REGULARIZATION METHODS FOR STATISTICAL INVERSE PROBLEMS AND APPLICATIONS

[11] Peter Mathé and Ulrich Tautenhahn, Regularization under general noise assumptions, Inverse Problems 27 (2011), no. 3,

Resolving the White Noise Paradox in the Regularisation of Inverse Problems

Model Selection and Geometry

Uncertainty principles for far field patterns and applications to inverse source problems

in Bounded Domains Ariane Trescases CMLA, ENS Cachan

Augmented Tikhonov Regularization

MCMC Sampling for Bayesian Inference using L1-type Priors

Institut für Numerische und Angewandte Mathematik

Preconditioned Newton methods for ill-posed problems

Two-parameter regularization method for determining the heat source

Morozov s discrepancy principle for Tikhonov-type functionals with non-linear operators

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

A Hybrid LSQR Regularization Parameter Estimation Algorithm for Large Scale Problems

A semismooth Newton method for L 1 data fitting with automatic choice of regularization parameters and noise calibration

An inverse scattering problem in random media

Spectral Regularization

= E [ξ ϕ ξ ψ ] satises 1

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

Regularization via Spectral Filtering

Regularization Theory

REGULARIZATION OF SOME LINEAR ILL-POSED PROBLEMS WITH DISCRETIZED RANDOM NOISY DATA

arxiv: v1 [math.st] 28 May 2016

Penalized Barycenters in the Wasserstein space

On Regularization Algorithms in Learning Theory

Inverse scattering problem from an impedance obstacle

Uncertainty Quantification for Inverse Problems. November 7, 2011

On Consistent Hypotheses Testing

On a Data Assimilation Method coupling Kalman Filtering, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model

Bayesian inverse problems with Laplacian noise

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution

SIGNAL AND IMAGE RESTORATION: SOLVING

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology

Sparse Approximation and Variable Selection

Linear convergence of iterative soft-thresholding

Class 2 & 3 Overfitting & Regularization

On regularization algorithms in learning theory

A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring

The impact of a curious type of smoothness conditions on convergence rates in l 1 -regularization

STAT 200C: High-dimensional Statistics

Statistically-Based Regularization Parameter Estimation for Large Scale Problems

Convex Optimization on Large-Scale Domains Given by Linear Minimization Oracles

Sequential Monte Carlo Samplers for Applications in High Dimensions

Anomalous transport of particles in Plasma physics

An example of Bayesian reasoning Consider the one-dimensional deconvolution problem with various degrees of prior information.

2 Tikhonov Regularization and ERM

High-dimensional regression with unknown variance

Convergence rates of the continuous regularized Gauss Newton method

Sparsity Regularization

ISyE 691 Data mining and analytics

Introduction to Bayesian methods in inverse problems

Conditional stability versus ill-posedness for operator equations with monotone operators in Hilbert space

arxiv: v1 [math.na] 6 Sep 2018

Compensation. June 29, Arizona State University. Edge Detection from Nonuniform Data via. Convolutional Gridding with Density.

ON THE DYNAMICAL SYSTEMS METHOD FOR SOLVING NONLINEAR EQUATIONS WITH MONOTONE OPERATORS

Kernels to detect abrupt changes in time series

Sparse Recovery in Inverse Problems

Direct estimation of linear functionals from indirect noisy observations

A Lower Bound Theorem. Lin Hu.

The Stein hull. Clément Marteau* Institut de Mathématiques, Université de Toulouse, INSA - 135, Avenue de Rangueil, F Toulouse Cedex 4, France

Provable Alternating Minimization Methods for Non-convex Optimization

Ill-Posedness of Backward Heat Conduction Problem 1

Numerische Mathematik

Nonlinear error dynamics for cycled data assimilation methods

Adaptive Filter Theory

Microlocal Methods in X-ray Tomography

COMPUTATION OF FOURIER TRANSFORMS FOR NOISY B FOR NOISY BANDLIMITED SIGNALS

New Statistical Model for the Enhancement of Noisy Speech

Minimum Message Length Analysis of the Behrens Fisher Problem

NON-EXTINCTION OF SOLUTIONS TO A FAST DIFFUSION SYSTEM WITH NONLOCAL SOURCES

Smoothness beyond dierentiability

A model function method in total least squares

Conditional Value-at-Risk (CVaR) Norm: Stochastic Case

Approximation Theoretical Questions for SVMs

Maximum Mean Discrepancy

Signal Denoising with Wavelets

Recovering overcomplete sparse representations from structured sensing

Modified Landweber iteration in Banach spaces convergence and convergence rates

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang

6.435, System Identification

Endogeneity in non separable models. Application to treatment models where the outcomes are durations

Bayesian Aggregation for Extraordinarily Large Dataset

A posteriori error estimates for the adaptivity technique for the Tikhonov functional and global convergence for a coefficient inverse problem

Bayesian Models for Regularization in Optimization

Inverse problems with Poisson data: Tikhonov-type regularization and iteratively regularized Newton methods

Inverse problem and optimization

Transcription:

Unbiased Risk Estimation as Parameter Choice Rule for Filter-based Regularization Methods Frank Werner 1 Statistical Inverse Problems in Biophysics Group Max Planck Institute for Biophysical Chemistry, Göttingen and Felix Bernstein Institute for Mathematical Statistics in the Biosciences University of Göttingen Chemnitz Symposium on Inverse Problems 2017 (on Tour in Rio) 1 joint work with Housen Li Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 1 / 34

Outline 1 Introduction 2 A posteriori parameter choice methods 3 Error analysis 4 Simulations 5 Conclusion Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 2 / 34

Introduction Outline 1 Introduction 2 A posteriori parameter choice methods 3 Error analysis 4 Simulations 5 Conclusion Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 3 / 34

Introduction Statistical inverse problems Setting: X, Y Hilbert spaces, T : X Y bounded, linear Task: Recover unknown f X from noisy measurements Y = Tf + σξ Noise: ξ is a standard Gaussian white noise process, σ > 0 noise level The model has to be understood in a weak sense: with ξ, g N Y g := Tf, g Y + σ ξ, g for all g Y ( ) 0, g 2 Y and E [ ξ, g 1 ξ, g 2 ] = g 1, g 2 Y. Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 4 / 34

Introduction Statistical inverse problems Assumptions: T is injective and Hilbert-Schmidt ( σ 2 k <, σ k singular values) σ is known exactly As the problem is ill-posed, regularization is needed. Consider filter-based regularization schemes ˆf α := q α (T T )T Y, α > 0. Aim: A posteriori choice of α such that rate of convergence (as σ 0) is order optimal (no loss of log-factors) Note: Heuristic parameter choice rules might work here as well, as the Bakushinskĭı veto does not hold in our setting (Becker 11). Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 5 / 34

A posteriori parameter choice methods Outline 1 Introduction 2 A posteriori parameter choice methods 3 Error analysis 4 Simulations 5 Conclusion Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 6 / 34

A posteriori parameter choice methods The discrepancy principle { For deterministic data: α DP = max α > 0 } T ˆf α Y τσ Y But here: Y / Y! Either pre-smoothing (Y Z := T Y X )...... or discretization: Y R n, ξ N n (0, I n ) and choose { α DP = max α > 0 T ˆfα Y τσ } n 2 Pros: Easy to implement Works for all q α Order-optimal convergence rates Cons: How to choose τ 1? Only discretized meaningful Early saturation Davies & Anderssen 86, Lukas 95, Blanchard, Hoffmann & Reiß 16 Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 7 / 34

A posteriori parameter choice methods The quasi-optimality criterion Neubauer 08 (r α (λ) = 1 λq α (λ)): α QO = argmin α>0 But for spectral cut-off r α (T T ) ˆf α = 0 for all α > 0 r α (T X T ) ˆf α Alternative formulation for Tikhonov regularization if candidates α 1 <... < α m are given: n QO = argmin ˆf αn ˆf X αn+1, α QO := α nqo. Pros: 1 n m 1 Easy to implement, very fast No knowledge of σ necessary Order-optimal convergence rates in mildly ill-posed situations Cons: Only for special q α Additional assumptions on noise and/or f necessary Performance unclear in severely ill-posed situations Bauer & Kindermann 08, Bauer & Reiß 08, Bauer & Kindermann 09 Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 8 / 34

A posteriori parameter choice methods The Lepskĭı-type balancing principle For given α, the standard deviation of ˆf α can be bounded by ( ) std (α) := σ Tr q αk (T T ) 2 T T If candidates α 1 <... < α m are given: { n LEP = max j } X ˆf αj ˆf αk 4κstd (α k ) for all 1 k j Pros: and α LEP = α nlep Works for all q α Robust in practice convergence rates (mildly / severely ill-posed) Cons: Computationally expansive κ 1 depends on decay of σ k loss of log factor compared to order-optimal rate Bauer & Pereverzev 05, Mathé 06, Mathé & Pereverzev 06 Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 9 / 34

A posteriori parameter choice methods Unbiased risk estimation Dating back to ideas of Mallows 73 and Stein 81 let ˆr (α, Y ) := T ˆf 2 α 2 T ˆf α, Y + 2σ 2 Tr (T Tq α (T T )) Y and choose α URE = argmin α>0 ˆr (α, Y ) [ T Note that E [ˆr (α, Y )] = E ˆf α Tf of α (Unbiased Risk Estimation) 2 Y ] c with c independent For spectral cut-off and in mildly ill-posed situations, this gives order optimal rates (Chernousova & Golubev 14). Besides this, only optimality in the image space is known (Li 87, Lukas 93, Kneip 94). Distributional behavior of α URE : Lucka et al 17. In general: Pros? Cons? Convergence rates? Order optimality? Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 10 / 34

Error analysis Outline 1 Introduction 2 A posteriori parameter choice methods 3 Error analysis 4 Simulations 5 Conclusion Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 11 / 34

Error analysis A priori parameter choice Assumptions Filter α q α (λ) C q and λ q α (λ) C q. Source condition W φ := { f X f = φ(t T )w, ω X C }. Note: for any f X there exists φ such that f W φ. Qualification condition The function φ is a qualification of q α if sup φ(λ) 1 λq α (λ) C φ φ(α) λ [0, T T Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 12 / 34

Error analysis A priori parameter choice Assumptions Let Σ (x) := # { k σ 2 k x} be the counting function of the singular values of T Approximation by smooth surrogate There exists S C 2, α 1 (0, T T ] and C S (0, 2) such that (1) lim α 0 S(α)/Σ(α) = 1 (approximation) (2) S < 0 (decreasing) (3) lim α S(α) = lim α S (α) = 0 (behavior above σ1 2) (4) lim α 0 αs(α) = 0 (Hilbert-Schmidt) (5) αs (α) is integrable on (0, α 1 ] S (6) (α) S (α) C S α on (0, α 1 ] Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 13 / 34

Error analysis A priori parameter choice A priori convergence rates Bissantz, Hohage, Munk, Ruymgaart 07 Let α φ(α ) 2 = σ 2 S(α ). (i) If φ is a qualification of q α, then [ ] ˆf α f 2 X sup E f W φ φ(α ) 2 = σ 2 S(α ) α as σ 0. (ii) If λ λφ(λ) is a qualification of the filter q α, then [ T ] sup E ˆf α Tf 2 α φ(α ) 2 = σ 2 S(α ) as σ 0. f W φ Y Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 14 / 34

Error analysis A priori parameter choice Mildly ill-posed situation: Example Assume σ 2 k k a, W b := Bissantz, Hohage, Munk, Ruymgaart 07 Let α (σ 2 ) a/(a+b+1). { f X : k=1 kb f 2 k 1 } with a > 1, b > 0: If φ (λ) = λ b/2a is a qualification of q α, then [ ] sup E ˆf α f 2 (σ 2 ) f W b X b a+b+1. If φ (λ) = λ b/2a+1/2 is a qualification of q α, then [ T ] sup E ˆfα Tf 2 (σ 2 ) a+b a+b+1. f W b Y These rates are order optimal over W b. Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 15 / 34

Error analysis Unbiased Risk estimation as parameter choice Unbiased risk estimation vs. the oracle Recall that ˆr (α, Y ) := 2 T ˆf α 2 T ˆf α, Y + 2σ 2 Tr (T Tq α (T T )) Y is an unbiased estimator for [ T r (α, f ) := E ˆfα Tf In the following, we will compare α URE = argmin α>0 2 Y ]. ˆr (α, Y ) and α o = argmin r(α, f ). α>0 Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 16 / 34

Error analysis Unbiased Risk estimation as parameter choice Additional assumptions (a) α {q α (σ 2 k )} k=1 is strictly monotone and continuous as R l2. (b) As α 0, αq α (α) c q > 0. (c) For α > 0, the function λ λq α (λ) is non-decreasing. Satisfied by Tikhonov, spectral cut-off, Landweber, iterated Tikhonov and Showalter regularization, under proper parametrization. E.g. Tikhonov with re-parametrization α α (q α (λ) = 1/( α + λ)) violates (b). ( λ ) (d) ψ (λ) := λφ 1 is convex (e) There exists a constant C q > c 2 q 1 ( Ψ (C q x) exp C for some explicitly known C > 0. such that ) x dx < with Ψ(x) := 2 x (S 1 (x)) 2. (d) can always be satisfied by weakening φ, (e) restricts the decay of the singular values Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 17 / 34

Error analysis Unbiased Risk estimation as parameter choice Oracle inequality Li & W. 16 There are positive constants C i, i = 1,..., 6, such that for all f W φ it holds [ ] E ˆfαURE f 2 C 1 ψ 1( C 2 r(α o, f ) + C 3 σ 2) + C 4 σ 2 X as σ 0. r(α o, f ) + σ r(α o, f ) + C 5 ( ) S 1 r(α C o,f ) 6 σ 2 Gives a comparison of the strong risk under α URE with the weak risk under the oracle α o. Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 18 / 34

Error analysis Unbiased Risk estimation as parameter choice Convergence rates Li & W. 16 If also λ λφ(λ) is a qualification of the filter q α, then for α φ(α ) 2 = σ 2 S(α ) there are C 1, C 2, C 3 > 0 such that as σ 0. [ ] sup E ˆf f 2 αure C 1 σ 2 S(α ) σ 2 S(α ) + C 2 f W φ X α S 1 (C 3 S(α )) If there is C 4 > 0 such that S(C 4 x) C 3 S(x), then this equals the a priori rate [ ] sup E ˆfαURE f 2 φ(α ) 2 = σ 2 S(α ). f W φ X α Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 19 / 34

Error analysis Unbiased Risk estimation as parameter choice Order optimality in mildly ill-posed situations Assume σ 2 k k a, W b := Oracle inequality For all f W: [ ] E ˆf f 2 αure X { f X : k=1 kb f 2 k 1 } with a > 1, b > 0: r(α o, f ) b a+b + σ 2a r(α o, f ) 1+a + σ 1 2a r(α o, f ) 1+2a 2. Convergence rate Thus, if λ λ b/2a+1/2 is a qualification of q α, then [ ] sup E ˆfαURE f 2 σ 2b a+b+1 f W b X which is order-optimal. Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 20 / 34

Error analysis Unbiased Risk estimation as parameter choice Unbiased risk estimation - pros and cons [ T 2 ] α URE = argmin ˆf α 2 T ˆf α, Y + 2σ 2 Tr (T Tq α (T T )) α>0 Y Pros: Works for many q α order-optimal convergence rates in mildly ill-posed situations no loss of log factor no tuning parameter Cons: Computationally expansive Early saturation performance in severely ill-posed situations unclear H. Li and F. Werner (2017). Empirical risk minimization as parameter choice rule for general linear regularization methods. Submitted, arxiv: 1703.07809. Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 21 / 34

Simulations Outline 1 Introduction 2 A posteriori parameter choice methods 3 Error analysis 4 Simulations 5 Conclusion Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 22 / 34

Simulations Rates of convergence A mildly ill-posed situation - antiderivative Let T : L 2 ([0, 1]) L 2 ([0, 1]) given by (Tf ) (x) = 1 0 min {x (1 y), y (1 x)} f (y) dy As (Tf ) = f the singular values σ k of T satisfy σ k k 2. We choose f (x) = Fourier coefficients: f k = ( 1)k 1 4π 3 k 2 ε > 0. { x if 0 x 1 2, 1 x if 1 2 x 1., so the optimal rate is O ( σ 3 4 ε) for any Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 23 / 34

Simulations Rates of convergence A mildly ill-posed situation - Tikhonov regularization 10 2 Empirical MISE 10 2 Empirical Variance 10 0 σ 3 4 10 4 10 2 10 10 10 4 10 7 10 5 10 3 10 1 10 16 10 7 10 5 10 3 10 1 Figure: Empirical MISE and variance of ˆf f 2 2 over 104 repetitions: α o ( ), α DP ( ), α QO ( ), α LEP ( ), α URE ( ). Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 24 / 34

Simulations Rates of convergence A severely ill-posed situation - satellite gradiometry Let R > 1 and S R 2 the unit sphere. Given g = 2 u r 2 u = 0 in R d \ B, u = f on S, ( u (x) = O x 1 2 ) as x 2. Corresponding T : L 2 (S, µ) L 2 (RS, µ) has singular values σ k = k ( k + 1) R k 2. on RS find f in We choose f (x) = π 2 x, x [ π, π] ( Optimal rate of convergence is O ( log (σ)) 3+ε) for any ε > 0. Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 25 / 34

Simulations Rates of convergence A severely ill-posed situation - Tikhonov regularization Empirical MISE 10 0 Empirical Variance 10 2 10 0 ( log (σ)) 3 10 4 10 8 10 2 10 8 10 6 10 4 10 3 10 2 10 12 10 8 10 6 10 4 10 3 10 2 Figure: Empirical MISE and variance of ˆf f 2 2 over 104 repetitions: α o ( ), α DP ( ), α QO ( ), α LEP ( ), α URE ( ). Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 26 / 34

Simulations Rates of convergence A severely ill-posed situation - backwards heat equation Let t > 0. Given g = u (, t) find f in u t (x, t) = 2 u (x, t) in ( π, π] (0, t), t 2 u (x, 0) = f (x) on [ π, π], u ( π, t) = u (π, t) on t (0, t]. Corresponding T : L 2 ([ π, π]) L 2 ([ π, π]) has singular values σ k = exp ( k 2 t ). We choose f (x) = π 2 x, x [ π, π] ( Optimal rate of convergence is O ( log (σ)) 3/2+ε) for any ε > 0. Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 27 / 34

Simulations Rates of convergence A severely ill-posed situation - Tikhonov regularization 10 1 Empirical MISE 10 2 Empirical Variance 10 0 ( log (σ)) 3 2 10 6 10 1 10 10 10 2 10 8 10 6 10 4 10 3 10 2 10 14 10 8 10 6 10 4 10 3 10 2 Figure: Empirical MISE and variance of ˆf f 2 2 over 104 repetitions: α o ( ), α DP ( ), α QO ( ), α LEP ( ), α URE ( ). Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 28 / 34

Simulations Efficiency simulations Efficiency simulations Measure the efficiency of a parameter choice rule α by the fraction [ ] E ˆfαo f 2 X R := [ ] E ˆf α f Numerical approximations of these as functions of σ with different parameters a, ν > 0 in the following setting: σ k = exp ( ak) f k = ±k ν (1 + N ( 0, 0.1 2)) Y k = σ k f k + N ( 0, σ 2) k = 1,..., 300, 10 4 repetitions Tikhonov regularization 2 X Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 29 / 34

Simulations Efficiency simulations Efficiency simulations - results 10 0 10 0 a = 0.2, ν = 1 a = 0.3, ν = 1 10 1 10 7 10 6 10 5 10 4 10 3 10 2 σ 10 1 10 7 10 6 10 5 10 4 10 3 10 2 σ Figure: R QO ( ), R DP ( ), R LEP ( ), and R URE ( ) Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 30 / 34

Simulations Efficiency simulations Efficiency simulations - results 10 0 10 0 a = 0.4, ν = 1 a = 0.6, ν = 1 10 1 10 7 10 6 10 5 10 4 10 3 10 2 σ 10 1 10 7 10 6 10 5 10 4 10 3 10 2 σ Figure: R QO ( ), R DP ( ), R LEP ( ), and R URE ( ) Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 31 / 34

Simulations Efficiency simulations Efficiency simulations - results 10 1 10 1 10 3 10 3 10 5 a = 0.3, ν = 3 10 5 a = 0.3, ν = 5 10 7 10 6 10 5 10 4 10 3 10 2 10 7 10 6 10 5 10 4 10 3 10 2 σ σ Figure: R QO ( ), R DP ( ), R LEP ( ), and R URE ( ) Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 32 / 34

Conclusion Outline 1 Introduction 2 A posteriori parameter choice methods 3 Error analysis 4 Simulations 5 Conclusion Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 33 / 34

Conclusion Presented results Analysis of a parameter choice based on unbiased risk estimation: oracle inequality convergence rates order optimality in mildly ill-posed situations Numerical comparison: in this specific setting, quasi-optimality outperforms all other methods unbiased risk estimation has higher variance (by design) simulations suggest order optimality of quasi-optimality also in severely ill-posed situations, not clear for unbiased risk estimation Thank you for your attention! Frank Werner, MPIbpC Göttingen Unbiased Risk Estimation October 30, 2017 34 / 34