SOTAVerified

Topological DeepONets and a generalization of the Chen-Chen operator approximation theorem

2026-03-12Unverified0· sign in to hype

Vugar Ismailov

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Deep Operator Networks (DeepONets) provide a branch-trunk neural architecture for approximating nonlinear operators acting between function spaces. In the classical operator approximation framework, the input is a function u C(K_1) defined on a compact set K_1 (typically a compact subset of a Banach space), and the operator maps u to an output function G(u) C(K_2) defined on a compact Euclidean domain K_2R^d. In this paper, we develop a topological extension in which the operator input lies in an arbitrary Hausdorff locally convex space X. We construct topological feedforward neural networks on X using continuous linear functionals from the dual space X^* and introduce topological DeepONets whose branch component acts on X through such linear measurements, while the trunk component acts on the Euclidean output domain. Our main theorem shows that continuous operators G:V C(K;R^m), where V X and KR^d are compact, can be uniformly approximated by such topological DeepONets. This extends the classical Chen-Chen operator approximation theorem from spaces of continuous functions to locally convex spaces and yields a branch-trunk approximation theorem beyond the Banach-space setting.

Reproductions