What Size Neural Network Gives Optimal Generalization? Convergence Properties of Backpropagation

dc.contributor.authorLawrence, Steveen_US
dc.contributor.authorGiles, C. Leeen_US
dc.contributor.authorTsoi, Ah Chungen_US
dc.date.accessioned2004-05-31T22:38:33Z
dc.date.available2004-05-31T22:38:33Z
dc.date.created1996-04en_US
dc.date.issued1998-10-15en_US
dc.description.abstractOne of the most important aspects of any machine learning paradigm is how it scales according to problem size and complexity. Using a task with known optimal training error, and a pre-specified maximum number of training updates, we investigate the convergence of the backpropagation algorithm with respect to a) the complexity of the required function approximation, b) the size of the network in relation to the size required for an optimal solution, and c) the degree of noise in the training data. In general, for a) the solution found is worse when the function to be approximated is more complex, for b) oversize networks can result in lower training and generalization error, and for c) the use of committee or ensemble techniques can be more beneficial as the amount of noise in the training data is increased. For the experiments we performed, we do not obtain the optimal solution in any case. We further support the observation that larger networks can produce better training and generalization error using a face recognition example where a network with many more parameters than training points generalizes better than smaller networks. (Also cross-referenced as UMIACS-TR-96-22)en_US
dc.format.extent2226577 bytes
dc.format.mimetypeapplication/postscript
dc.identifier.urihttp://hdl.handle.net/1903/809
dc.language.isoen_US
dc.relation.isAvailableAtDigital Repository at the University of Marylanden_US
dc.relation.isAvailableAtUniversity of Maryland (College Park, Md.)en_US
dc.relation.isAvailableAtTech Reports in Computer Science and Engineeringen_US
dc.relation.isAvailableAtUMIACS Technical Reportsen_US
dc.relation.ispartofseriesUM Computer Science Department; CS-TR-3617en_US
dc.relation.ispartofseriesUMIACS; UMIACS-TR-96-22en_US
dc.titleWhat Size Neural Network Gives Optimal Generalization? Convergence Properties of Backpropagationen_US
dc.typeTechnical Reporten_US

Files

Original bundle

Now showing 1 - 2 of 2
No Thumbnail Available
Name:
CS-TR-3617.ps
Size:
2.12 MB
Format:
Postscript Files
Loading...
Thumbnail Image
Name:
CS-TR-3617.pdf
Size:
575.02 KB
Format:
Adobe Portable Document Format
Description:
Auto-generated copy of CS-TR-3617.ps