Extracting Symbolic Representations Learned by Neural Networks

dc.contributor.advisorReggia, James Aen_US
dc.contributor.authorHuynh, Thuan Quangen_US
dc.contributor.departmentComputer Scienceen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2012-07-10T05:45:02Z
dc.date.available2012-07-10T05:45:02Z
dc.date.issued2012en_US
dc.description.abstractUnderstanding what neural networks learn from training data is of great interest in data mining, data analysis, and critical applications, and in evaluating neural network models. Unfortunately, the product of neural network training is typically opaque matrices of floating point numbers that are not obviously understandable. This difficulty has inspired substantial past research on how to extract symbolic, human-readable representations from a trained neural network, but the results obtained so far are very limited (e.g., large rule sets produced). This problem occurs in part due to the distributed hidden layer representation created during learning. Most past symbolic knowledge extraction algorithms have focused on progressively more sophisticated ways to cluster this distributed representation. In contrast, in this dissertation, I take a different approach. I develop ways to alter the error backpropagation neural network training process itself so that it creates a representation of what has been learned in the hidden layer activation space that is more amenable to existing symbolic representation extraction methods. In this context, this dissertation research makes four main contributions. First, modifications to the backpropagation learning procedure are derived mathematically, and it is shown that these modifications can be accomplished as local computations. Second, the effectiveness of the modified learning procedure for feedforward networks is established by showing that, on a set of benchmark tasks, it produces rule sets that are substantially simpler than those produced by standard backpropagation learning. Third, this approach is extended to simple recurrent networks, and experimental evaluation shows remarkable reduction in the sizes of the finite state machines extracted from the recurrent networks trained using this approach. Finally, this method is further modified to work on echo state networks, and computational experiments again show significant improvement in finite state machine extraction from these networks. These results clearly establish that principled modification of error backpropagation so that it constructs a better separated hidden layer representation is an effective way to improve contemporary symbolic extraction methods.en_US
dc.identifier.urihttp://hdl.handle.net/1903/12841
dc.subject.pqcontrolledComputer scienceen_US
dc.subject.pquncontrolledhidden layer representationen_US
dc.subject.pquncontrolledneural networken_US
dc.subject.pquncontrolledpenalty functionen_US
dc.subject.pquncontrolledrule extractionen_US
dc.titleExtracting Symbolic Representations Learned by Neural Networksen_US
dc.typeDissertationen_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Huynh_umd_0117E_13271.pdf
Size:
1.52 MB
Format:
Adobe Portable Document Format