TY - JOUR
T1 - Consistency analysis of bilevel data-driven learning in inverse problems
AU - Chada, Neil Kumar
AU - Schillings, Claudia
AU - Tong, Xin T.
AU - Weissmann, Simon
N1 - KAUST Repository Item: Exported on 2021-12-13
Acknowledgements: NKC acknowledges a Singapore Ministry of Education Academic Research Funds Tier 2 grant [MOE2016-T2-2-135] and KAUST baseline funding. SW is grateful to the DFG RTG1953 “Statistical Modeling of Complex Systems and Processes” for funding of this research. The research of XTT is supported by the National University of Singapore grant R-146-000-292-114.
PY - 2021/12/10
Y1 - 2021/12/10
N2 - One fundamental problem when solving inverse problems is how to find regularization parameters. This article considers solving this problem using data-driven bilevel optimization, i.e. we consider the adaptive learning of the regularization parameter from data by means of optimization. This approach can be interpreted as solving an empirical risk minimization problem, and we analyze its performance in the large data sample size limit for general nonlinear problems. We demonstrate how to implement our framework on linear inverse problems, where we can further show that the inverse accuracy does not depend on the ambient space dimension. To reduce the associated computational cost, online numerical schemes are derived using the stochastic gradient descent method. We prove convergence of these numerical schemes under suitable assumptions on the forward problem. Numerical experiments are presented illustrating the theoretical results and demonstrating the applicability and efficiency of the proposed approaches for various linear and nonlinear inverse problems, including Darcy flow, the eikonal equation, and an image denoising example.
AB - One fundamental problem when solving inverse problems is how to find regularization parameters. This article considers solving this problem using data-driven bilevel optimization, i.e. we consider the adaptive learning of the regularization parameter from data by means of optimization. This approach can be interpreted as solving an empirical risk minimization problem, and we analyze its performance in the large data sample size limit for general nonlinear problems. We demonstrate how to implement our framework on linear inverse problems, where we can further show that the inverse accuracy does not depend on the ambient space dimension. To reduce the associated computational cost, online numerical schemes are derived using the stochastic gradient descent method. We prove convergence of these numerical schemes under suitable assumptions on the forward problem. Numerical experiments are presented illustrating the theoretical results and demonstrating the applicability and efficiency of the proposed approaches for various linear and nonlinear inverse problems, including Darcy flow, the eikonal equation, and an image denoising example.
UR - http://hdl.handle.net/10754/673987
UR - https://www.intlpress.com/site/pub/pages/journals/items/cms/content/vols/0020/0001/a004/
U2 - 10.4310/cms.2022.v20.n1.a4
DO - 10.4310/cms.2022.v20.n1.a4
M3 - Article
SN - 1539-6746
VL - 20
SP - 123
EP - 164
JO - Communications in Mathematical Sciences
JF - Communications in Mathematical Sciences
IS - 1
ER -