Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited

Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the st...

Full description

Bibliographic Details
Main Authors: Sarah E. Marzen, James P. Crutchfield
Format: Article
Language:English
Published: MDPI AG 2022-01-01
Series:Entropy
Subjects:
Online Access:https://www.mdpi.com/1099-4300/24/1/90
_version_ 1797494235603140608
author Sarah E. Marzen
James P. Crutchfield
author_facet Sarah E. Marzen
James P. Crutchfield
author_sort Sarah E. Marzen
collection DOAJ
description Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the stochastic processes generated by a large suite of probabilistic deterministic finite-state automata (PDFA) in the small-data limit according to two metrics: predictive accuracy and distance to a predictive rate-distortion curve. The latter provides a sense of whether or not the RNN is a lossy predictive feature extractor in the information-theoretic sense. PDFAs provide an excellent performance benchmark in that they can be systematically enumerated, the randomness and correlation structure of their generated processes are exactly known, and their optimal memory-limited predictors are easily computed. With less data than is needed to make a good prediction, LSTMs surprisingly lose at predictive accuracy, but win at lossy predictive feature extraction. These results highlight the utility of causal states in understanding the capabilities of RNNs to predict.
first_indexed 2024-03-10T01:31:26Z
format Article
id doaj.art-30f2d8626f49497289f0fce13ef75856
institution Directory Open Access Journal
issn 1099-4300
language English
last_indexed 2024-03-10T01:31:26Z
publishDate 2022-01-01
publisher MDPI AG
record_format Article
series Entropy
spelling doaj.art-30f2d8626f49497289f0fce13ef758562023-11-23T13:41:44ZengMDPI AGEntropy1099-43002022-01-012419010.3390/e24010090Probabilistic Deterministic Finite Automata and Recurrent Networks, RevisitedSarah E. Marzen0James P. Crutchfield1W. M. Keck Science Department, Pitzer, Scripps, and Claremont McKenna College, Claremont, CA 91711, USAComplexity Sciences Center, Physics Department, University of California at Davis, One Shields Avenue, Davis, CA 95616, USAReservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the stochastic processes generated by a large suite of probabilistic deterministic finite-state automata (PDFA) in the small-data limit according to two metrics: predictive accuracy and distance to a predictive rate-distortion curve. The latter provides a sense of whether or not the RNN is a lossy predictive feature extractor in the information-theoretic sense. PDFAs provide an excellent performance benchmark in that they can be systematically enumerated, the randomness and correlation structure of their generated processes are exactly known, and their optimal memory-limited predictors are easily computed. With less data than is needed to make a good prediction, LSTMs surprisingly lose at predictive accuracy, but win at lossy predictive feature extraction. These results highlight the utility of causal states in understanding the capabilities of RNNs to predict.https://www.mdpi.com/1099-4300/24/1/90time series predictionfinite state machineshidden Markov modelsrecurrent neural networksreservoir computerslong short-term memory
spellingShingle Sarah E. Marzen
James P. Crutchfield
Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
Entropy
time series prediction
finite state machines
hidden Markov models
recurrent neural networks
reservoir computers
long short-term memory
title Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_full Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_fullStr Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_full_unstemmed Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_short Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_sort probabilistic deterministic finite automata and recurrent networks revisited
topic time series prediction
finite state machines
hidden Markov models
recurrent neural networks
reservoir computers
long short-term memory
url https://www.mdpi.com/1099-4300/24/1/90
work_keys_str_mv AT sarahemarzen probabilisticdeterministicfiniteautomataandrecurrentnetworksrevisited
AT jamespcrutchfield probabilisticdeterministicfiniteautomataandrecurrentnetworksrevisited