How Embedded Memory in Recurrent Neural Network Architectures Helps Learning Long-term Dependencies

dc.contributor.authorLin, Tsungnanen_US
dc.contributor.authorHorne, Bill G.en_US
dc.contributor.authorGiles, C. Leeen_US
dc.date.accessioned2004-05-31T22:38:52Z
dc.date.available2004-05-31T22:38:52Z
dc.date.created1996-08en_US
dc.date.issued1998-10-15en_US
dc.description.abstractLearning long-term temporal dependencies with recurrent neural networks can be a difficult problem. It has recently been shown that a class of recurrent neural networks called NARX networks perform much better than conventional recurrent neural networks for learning certain simple long-term dependency problems. The intuitive explanation for this behavior is that the output memories of a NARX network can be manifested as jump-ahead connections in the time-unfolded network. These jump-ahead connections can propagate gradient information more efficiently, thus reducing the sensitivity of the network to long-term dependencies. This work gives empirical justification to our hypothesis that similar improvements in learning long-term dependencies can be achieved with other classes of recurrent neural network architectures simply by increasing the order of the embedded memory. In particular we explore the impact of learning simple long-term dependency problems on three classes of recurrent neural networks architectures: globally recurrent networks, locally recurrent networks, and NARX (output feedback) networks. Comparing the performance of these architectures with different orders of embedded memory on two simple long-term dependences problems shows that all of these classes of networks architectures demonstrate significant improvement on learning long-term dependencies when the orders of embedded memory are increased. These results can be important to a user comfortable to a specific recurrent neural network architecture because simply increasing the embedding memory order will make the architecture more robust to the problem of long-term dependency learning. (Also cross-referenced as UMIACS-TR-96-28)en_US
dc.format.extent275595 bytes
dc.format.mimetypeapplication/postscript
dc.identifier.urihttp://hdl.handle.net/1903/813
dc.language.isoen_US
dc.relation.isAvailableAtDigital Repository at the University of Marylanden_US
dc.relation.isAvailableAtUniversity of Maryland (College Park, Md.)en_US
dc.relation.isAvailableAtTech Reports in Computer Science and Engineeringen_US
dc.relation.isAvailableAtUMIACS Technical Reportsen_US
dc.relation.ispartofseriesUM Computer Science Department; CS-TR-3626en_US
dc.relation.ispartofseriesUMIACS; UMIACS-TR-96-28en_US
dc.titleHow Embedded Memory in Recurrent Neural Network Architectures Helps Learning Long-term Dependenciesen_US
dc.typeTechnical Reporten_US

Files

Original bundle
Now showing 1 - 2 of 2
No Thumbnail Available
Name:
CS-TR-3626.ps
Size:
269.14 KB
Format:
Postscript Files
Loading...
Thumbnail Image
Name:
CS-TR-3626.pdf
Size:
181.41 KB
Format:
Adobe Portable Document Format
Description:
Auto-generated copy of CS-TR-3626.ps