Neural Networks as Universal Finite-State Machines: A Constructive Deterministic Finite Automaton Theory
Neural Networks as Universal Finite-State Machines: A Constructive Deterministic Finite Automaton Theory
We present a complete theoretical and empirical framework establishing feedforward neural networks as universal finite-state machines (N-FSMs). Our results prove that finite-depth ReLU and threshold networks can exactly simulate deterministic finite automata (DFAs) by unrolling state transitions into depth-wise neural layers, with formal characterizations of required depth, width, and state compression. We demonstrate that DFA transitions are linearly separable, binary threshold activations allow exponential compression, and Myhill-Nerode equivalence classes can be embedded into continuous latent spaces while preserving separability. We also formalize the expressivity boundary: fixed-depth feedforward networks cannot recognize non-regular languages requiring unbounded memory. Unlike prior heuristic or probing-based studies, we provide constructive proofs and design explicit DFA-unrolled neural architectures that empirically validate every claim. Our results bridge deep learning, automata theory, and neural-symbolic computation, offering a rigorous blueprint for how discrete symbolic processes can be realized in continuous neural systems.
Sahil Rajesh Dhayalkar
自动化基础理论计算技术、计算机技术
Sahil Rajesh Dhayalkar.Neural Networks as Universal Finite-State Machines: A Constructive Deterministic Finite Automaton Theory[EB/OL].(2025-05-16)[2025-06-08].https://arxiv.org/abs/2505.11694.点此复制
评论