Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited.

Entropy (Basel)

Complexity Sciences Center, Physics Department, University of California at Davis, One Shields Avenue, Davis, CA 95616, USA.

Published: January 2022

AI Article Synopsis

  • Reservoir computers (RCs) and recurrent neural networks (RNNs) theoretically can replicate any finite-state automaton, and this study tests their real-life effectiveness using generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNNs against a set of probabilistic deterministic finite-state automata (PDFA).
  • The analysis focuses on two main metrics: predictive accuracy and how close the predictions are to a theoretical optimal predictive rate-distortion curve, which helps determine if the RNN acts as a lossy predictive feature extractor.
  • While LSTMs struggle with predictive accuracy in small-data scenarios, they excel in extracting useful features, indicating the importance of understanding causal states in evaluating RNN capabilities.

Article Abstract

Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the stochastic processes generated by a large suite of probabilistic deterministic finite-state automata (PDFA) in the small-data limit according to two metrics: predictive accuracy and distance to a predictive rate-distortion curve. The latter provides a sense of whether or not the RNN is a lossy predictive feature extractor in the information-theoretic sense. PDFAs provide an excellent performance benchmark in that they can be systematically enumerated, the randomness and correlation structure of their generated processes are exactly known, and their optimal memory-limited predictors are easily computed. With less data than is needed to make a good prediction, LSTMs surprisingly lose at predictive accuracy, but win at lossy predictive feature extraction. These results highlight the utility of causal states in understanding the capabilities of RNNs to predict.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774624PMC
http://dx.doi.org/10.3390/e24010090DOI Listing

Publication Analysis

Top Keywords

probabilistic deterministic
8
predictive accuracy
8
lossy predictive
8
predictive feature
8
predictive
5
deterministic finite
4
finite automata
4
automata recurrent
4
recurrent networks
4
networks revisited
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!