Adaptive Choice of the Regularization Parameter in Numerical Differentiation
Abstract
We investigate a novel adaptive choice rule of the Tikhonov regularization parameter in numerical differentiation which is a classic ill-posed problem. By assuming a general unknown Hölder type error estimate derived for numerical differentiation, we choose a regularization parameter in a geometric set providing a nearly optimal convergence rate with very limited a-priori information. Numerical simulation in image edge detection verifies reliability and efficiency of the new adaptive approach.
About this article
How to Cite
Adaptive Choice of the Regularization Parameter in Numerical Differentiation. (2018). Journal of Computational Mathematics, 33(4), 415-427. https://doi.org/10.4208/jcm.1503-m2014-0134