Deep Thinking Systems: Logical Extrapolation with Recurrent Neural Networks

dc.contributor.advisorGoldstein, Tomen_US
dc.contributor.authorSchwarzschild, Avi Koplonen_US
dc.contributor.departmentMathematicsen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2023-06-26T05:38:05Z
dc.date.available2023-06-26T05:38:05Z
dc.date.issued2023en_US
dc.description.abstractDeep neural networks are powerful machines for visual pattern recognition, but reasoning tasks that are easy for humans are still be difficult for neural models. Humans possess the ability to extrapolate reasoning strategies learned on simple problems to solve harder examples, often by thinking for longer. We study neural networks that have exactly this capability. By employing recurrence, we build neural networks that can expend more computation when needed. Using several datasets designed specifically for studying generalization from easy problems to harder test samples, we show that our recurrent networks can extrapolate from easy training data to much harder examples at test time, and they do so with many more iterations of a recurrent block of layers than are used during training.en_US
dc.identifierhttps://doi.org/10.13016/dspace/yi8j-hqys
dc.identifier.urihttp://hdl.handle.net/1903/30189
dc.language.isoenen_US
dc.subject.pqcontrolledArtificial intelligenceen_US
dc.subject.pqcontrolledComputer scienceen_US
dc.titleDeep Thinking Systems: Logical Extrapolation with Recurrent Neural Networksen_US
dc.typeDissertationen_US

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
Schwarzschild_umd_0117E_23303.pdf
Size:
14.29 MB
Format:
Adobe Portable Document Format