Computational Capabilities of Recurrent NARX Neural Networks
View/ Open
Date
1998-10-15Author
Siegelmann, Hava T.
Horne, Bill G.
Giles, C. Lee
Metadata
Show full item recordAbstract
Recently, fully connected recurrent neural networks have been proven to be
computationally rich --- at least as powerful as Turing machines. This
work focuses on another network which is popular in control applications
and has been found to be very effective at learning a variety of problems.
These networks are based upon Nonlinear AutoRegressive models with
eXogenous Inputs (NARX models), and are therefore called {\em NARX
networks}. As opposed to other recurrent networks, NARX networks have a
limited feedback which comes only from the output neuron rather than from
hidden states. They are formalized by \[ y(t) = \Psi \left(
\rule[-1ex]{0em}{3ex} u(t-n_u), \ldots, u(t-1), u(t), y(t-n_y), \ldots,
y(t-1) \right), \] where $u(t)$ and $y(t)$ represent input and output of
the network at time $t$, $n_u$ and $n_y$ are the input and output order,
and the function $\Psi$ is the mapping performed by a Multilayer
Perceptron. We constructively prove that the NARX networks with a finite
number of parameters are computationally as strong as fully connected
recurrent networks and thus Turing machines. We conclude that in theory
one can use the NARX models, rather than conventional recurrent networks
without any computational loss even though their feedback is limited.
Furthermore, these results raise the issue of what amount of feedback or
recurrence is necessary for any network to be Turing equivalent and what
restrictions on feedback limit computational power.
(Also cross-referenced as UMIACS-TR-95-12)