Variational Regularization Theory Based on Image Space Approximation Rates

09/01/2020
by   Philip Miller, et al.
0

We present a new approach to convergence rate results for variational regularization. Avoiding Bregman distances and using image space approximation rates as source conditions we prove a nearly minimax theorem showing that the modulus of continuity is an upper bound on the reconstruction error up to a constant. Applied to Besov space regularization we obtain convergence rate results for 0,2,q- and 0,p,p-penalties without restrictions on p,q∈ (1,∞). Finally we prove equivalence of Hölder-type variational source conditions, bounds on the defect of the Tikhonov functional, and image space approximation rates.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/07/2017

Convergence Rates of Variational Posterior Distributions

We study convergence rates of variational posterior distributions for no...
research
08/06/2020

Exact Convergence Rate Analysis of the Independent Metropolis-Hastings Algorithms

A well-known difficult problem regarding Metropolis-Hastings algorithms ...
research
05/03/2020

Variational Shape Approximation of Point Set Surfaces

In this work, we present a translation of the complete pipeline for vari...
research
09/20/2023

Convergence rate of numerical scheme for SDEs with a distributional drift in Besov space

This paper is concerned with numerical solutions of one-dimensional SDEs...
research
12/20/2020

On Tikhonov functionals penalized by Bregman distances

We investigate Tikhonov regularization methods for nonlinear ill-posed p...
research
07/09/2020

Geometric Bounds for Convergence Rates of Averaging Algorithms

We develop a generic method for bounding the convergence rate of an aver...
research
05/29/2020

Maximal Spaces for Approximation Rates in ℓ^1-regularization

We study Tikhonov regularization for possibly nonlinear inverse problems...

Please sign up or login with your details

Forgot password? Click here to reset