Some density theorems in neural network with variable exponent
Some density theorems in neural network with variable exponent
In this paper, we extend several approximation theorems, originally formulated in the context of the standard $L^p$ norm, to the more general framework of variable exponent spaces. Our study is motivated by applications in neural networks, where function approximation plays a crucial role. In addition to these generalizations, we provide alternative proofs for certain well-known results concerning the universal approximation property. In particular, we highlight spaces with variable exponents as illustrative examples, demonstrating the broader applicability of our approach.
Mitsuo Izuki、Takahiro Noi、Yoshihiro Sawano、Hirokazu Tanaka
数学计算技术、计算机技术
Mitsuo Izuki,Takahiro Noi,Yoshihiro Sawano,Hirokazu Tanaka.Some density theorems in neural network with variable exponent[EB/OL].(2025-04-19)[2025-05-02].https://arxiv.org/abs/2504.14476.点此复制
评论