Revealing structure in trained neural networks through dimensionality-based methods

dc.contributor.advisorShea-Brown, Eric
dc.contributor.authorFarrell, Matthew Stuart
dc.date.accessioned2020-10-26T20:39:03Z
dc.date.available2020-10-26T20:39:03Z
dc.date.issued2020-10-26
dc.date.submitted2020
dc.descriptionThesis (Ph.D.)--University of Washington, 2020
dc.description.abstractNeural networks trained by machine learning optimization methods are currently being analyzed to shed light on brain function. While exciting progress is being made, the complicated nature of the network models typically considered has made ``opening the black box'' a significant challenge. In this thesis I approach the problem by starting with network models and tasks that can understood more easily, but that capture fundamental elements of more complex models. I reveal new aspects of the behavior of these models through the lens of effective dimensionality, which quantifies the number of axes needed to describe data. Through this investigation a new idea of "dimensionality balance'' emerges, where neural networks trained with stochastic gradient descent automatically strike a balance between increasing dimensionality (to more easily distinguish between different objects) and decreasing dimensionality (to build invariance to different examples of the same object). Mathematical analysis reveals the core mechanisms that may underlie the effects, and experiments with the image classification network VGG indicate that this balance is a general phenomenon. Finally, I demonstrate how and why dimensionality reduction methods can be used to extract information from network weights in a simple model, laying some guiding principles for ways of extracting insights from the recent explosion of brain connectomics data.
dc.embargo.termsOpen Access
dc.format.mimetypeapplication/pdf
dc.identifier.otherFarrell_washington_0250E_22191.pdf
dc.identifier.urihttp://hdl.handle.net/1773/46368
dc.language.isoen_US
dc.rightsCC BY
dc.subjectartificial neural networks
dc.subjectconnectomics
dc.subjectdeep learning
dc.subjectdimensionality
dc.subjectrepresentation learning
dc.subjectNeurosciences
dc.subjectArtificial intelligence
dc.subject.otherApplied mathematics
dc.titleRevealing structure in trained neural networks through dimensionality-based methods
dc.typeThesis

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Farrell_washington_0250E_22191.pdf
Size:
32.21 MB
Format:
Adobe Portable Document Format