Learning Long-Term Dependencies is Not as Difficult With NARX Recurrent Neural Networks

dc.contributor.authorLin, Tsungnanen_US
dc.contributor.authorHorne, Bill G.en_US
dc.contributor.authorTino, Peteren_US
dc.contributor.authorGiles, C. Leeen_US
dc.date.accessioned2004-05-31T22:33:43Z
dc.date.available2004-05-31T22:33:43Z
dc.date.created1995-07en_US
dc.date.issued1998-10-15en_US
dc.description.abstractIt has recently been shown that gradient descent learning algorithms for recurrent neural networks can perform poorly on tasks that involve long- term dependencies, i.e. those problems for which the desired output depends on inputs presented at times far in the past. In this paper we explore the long-term dependencies problem for a class of architectures called NARX recurrent neural networks, which have power ful representational capabilities. We have previously reported that gradient descent learning is more effective in NARX networks than in recurrent neural network architectures that have ``hidden states'' on problems includ ing grammatical inference and nonlinear system identification. Typically, the network converges much faster and generalizes better than other net works. The results in this paper are an attempt to explain this phenomenon. We present some experimental results which show that NARX networks can often retain information for two to three times as long as conventional recurrent neural networks. We show that although NARX networks do not circumvent the problem of long-term dependencies, they can greatly improve performance on long-term dependency problems. We also describe in detail some of the assumption regarding what it means to latch information robustly and suggest possible ways to loosen these assumptions. (Also cross-referenced as UMIACS-TR-95-78)en_US
dc.format.extent481301 bytes
dc.format.mimetypeapplication/postscript
dc.identifier.urihttp://hdl.handle.net/1903/745
dc.language.isoen_US
dc.relation.isAvailableAtDigital Repository at the University of Marylanden_US
dc.relation.isAvailableAtUniversity of Maryland (College Park, Md.)en_US
dc.relation.isAvailableAtTech Reports in Computer Science and Engineeringen_US
dc.relation.isAvailableAtUMIACS Technical Reportsen_US
dc.relation.ispartofseriesUM Computer Science Department; CS-TR-3500en_US
dc.relation.ispartofseriesUMIACS; UMIACS-TR-95-78en_US
dc.titleLearning Long-Term Dependencies is Not as Difficult With NARX Recurrent Neural Networksen_US
dc.typeTechnical Reporten_US

Files

Original bundle
Now showing 1 - 2 of 2
No Thumbnail Available
Name:
CS-TR-3500.ps
Size:
470.02 KB
Format:
Postscript Files
Loading...
Thumbnail Image
Name:
CS-TR-3500.pdf
Size:
226.58 KB
Format:
Adobe Portable Document Format
Description:
Auto-generated copy of CS-TR-3500.ps