Document Type

Article

Publication Date

2021

Publication Title

Journal of Machine Learning Research

Volume

22

Pages

225 (1-65)

Abstract

Learning a function from a finite number of sampled data points (measurements) is a fundamental problem in science and engineering. This is often formulated as a minimum norm interpolation (MNI) problem, a regularized learning problem or, in general, a semi discrete inverse problem (SDIP), in either Hilbert spaces or Banach spaces. The goal of this paper is to systematically study solutions of these problems in Banach spaces. We aim at obtaining explicit representer theorems for their solutions, on which convenient solution methods can then be developed. For the MNI problem, the explicit representer theorems enable us to express the infimum in terms of the norm of the linear combination of the interpolation functionals. For the purpose of developing efficient computational algorithms, we establish the fixed-point equation formulation of solutions of these problems. We reveal that unlike in a Hilbert space, in general, solutions of these problems in a Banach space may not be able to be reduced to truly finite dimensional problems (with certain infinite dimensional components hidden). We demonstrate how this obstacle can be removed, reducing the original problem to a truly finite dimensional one, in the special case when the Banach space is ℓ1(N).

Comments

© 2021 Rui Wang and Yuesheng Xu.

Published under a Creative Commons Attribution 4.0 International (CC BY 4.0) license.

Original Publication Citation

Wang, R., & Xu, Y. (2021). Representer theorems in Banach spaces: Minimum norm interpolation, regularized learning and semi-discrete inverse problems. Journal of Machine Learning Research, 22, 1-65, Article 225. https://jmlr.org/papers/v22/20-751.html

Share

COinS