Learning a Class of Large Finite State Machines with a Recurrent Neural Network

dc.contributor.authorGiles, C. Leeen_US
dc.contributor.authorHorne, Bill G.en_US
dc.contributor.authorLin, T.en_US
dc.date.accessioned2004-05-31T22:27:36Z
dc.date.available2004-05-31T22:27:36Z
dc.date.created1994-08en_US
dc.date.issued1998-10-15en_US
dc.description.abstractOne of the issues in any learning model is how it scales with problem size. Neural networks have not been immune to scaling issues. We show that a dynamically-driven discrete-time recurrent network (DRNN) can learn rather large grammatical inference problems when the strings of a finite memory machine (FMM) are encoded as temporal sequences. FMMs are a subclass of finite state machines which have a finite memory or a finite order of inputs and outputs. The DRNN that learns the FMM is a neural network that maps directly from the sequential machine implementation of the FMM. It has feedback only from the output and not from any hidden units; an example is the recurrent network of Narendra and Parthasarathy. (FMMs that have zero order in the feedback of outputs are called definite memory machines and are analogous to Time-delay or Finite Impulse Response neural networks.) Due to their topology these DRNNs are as least as powerful as any sequential machine implementation of a FMM and should be capable of representing any FMM. We choose to learn ``particular FMMs.\' Specifically, these FMMs have a large number of states (simulations are for $256$ and $512$ state FMMs) but have minimal order, relatively small depth and little logic when the FMM is implemented as a sequential machine. Simulations for the number of training examples versus generalization performance and FMM extraction size show that the number of training samples necessary for perfect generalization is less than that necessary to completely characterize the FMM to be learned. This is in a sense a best case learning problem since any arbitrarily chosen FMM with a minimal number of states would have much more order and string depth and most likely require more logic in its sequential machine implementation. (Also cross-referenced as UMIACS-TR-94-94)en_US
dc.format.extent448314 bytes
dc.format.mimetypeapplication/postscript
dc.identifier.urihttp://hdl.handle.net/1903/657
dc.language.isoen_US
dc.relation.isAvailableAtDigital Repository at the University of Marylanden_US
dc.relation.isAvailableAtUniversity of Maryland (College Park, Md.)en_US
dc.relation.isAvailableAtTech Reports in Computer Science and Engineeringen_US
dc.relation.isAvailableAtUMIACS Technical Reportsen_US
dc.relation.ispartofseriesUM Computer Science Department; CS-TR-3328en_US
dc.relation.ispartofseriesUMIACS; UMIACS-TR-94-94en_US
dc.titleLearning a Class of Large Finite State Machines with a Recurrent Neural Networken_US
dc.typeTechnical Reporten_US

Files

Original bundle
Now showing 1 - 2 of 2
No Thumbnail Available
Name:
CS-TR-3328.ps
Size:
437.81 KB
Format:
Postscript Files
Loading...
Thumbnail Image
Name:
CS-TR-3328.pdf
Size:
338.7 KB
Format:
Adobe Portable Document Format
Description:
Auto-generated copy of CS-TR-3328.ps