Deep learning frameworks for modeling how neural circuits learn

dc.contributor.advisorShea-Brown, Eric
dc.contributor.authorLiu, Yuhan Helena
dc.date.accessioned2024-09-09T23:03:12Z
dc.date.available2024-09-09T23:03:12Z
dc.date.issued2024-09-09
dc.date.submitted2024
dc.descriptionThesis (Ph.D.)--University of Washington, 2024
dc.description.abstractThe brain's prowess in learning and adapting remains an enigma, particularly in its approach to the 'temporal credit assignment' problem. How do neural circuits determine which specific states and connections contribute to future outcomes, and subsequently adjust these for enhanced learning? My thesis addresses this by combining insights from the latest large-scale neuroscience data and recent deep learning theoretical tools. The first two projects introduce novel learning rules inspired by the Allen Institute's transcriptomics data, which revealed widespread and intricate cell-type-specific interactions among neuromodulatory molecules. This rule enables neurons to propagate credit information efficiently, enhancing learning performance beyond that of biologically plausible predecessors. Extensive computational experiments confirm the significant role of local neuromodulatory signals in learning, offering new perspectives on neural information processing. My third project assesses the generalization capabilities of bio-plausible learning rules through the lens of deep learning theory, particularly focusing on the curvature of the loss landscape via the loss’ Hessian eigenspectrum. Our findings reveal that these rules often settle in high-curvature regions of the loss landscape, indicating suboptimal generalization. This analysis led to a mathematical theorem linking synaptic weight update dynamics to landscape curvature, proposing neuromodulator-driven adjustments as a potential enhancement for learning rule performance. Given how initial conditions can greatly influence a system’s future trajectory, the fourth project delves into the impact of initial connectivity structures on learning dynamics in neural circuits. By examining various connectivity patterns derived from neuroscience data, including recent electron microscopy data, we analyze how these structures influence learning regimes, implicating metabolic costs and risks of catastrophic forgetting. Our findings suggest that high-rank initializations utilize pre-existing high-dimensional input expansion to facilitate input decoding, leading to minimal changes post-training and increasing the propensity for lazy learning. These specific initializations thus predispose networks toward certain learning behaviors, critically affecting their ability to adapt and generalize.
dc.embargo.termsOpen Access
dc.format.mimetypeapplication/pdf
dc.identifier.otherLiu_washington_0250E_26903.pdf
dc.identifier.urihttps://hdl.handle.net/1773/51782
dc.language.isoen_US
dc.rightsCC BY-NC-SA
dc.subjectArtificial intelligence
dc.subjectComputational neuroscience
dc.subjectDeep learning theory
dc.subjectLearning and plasticity
dc.subjectNeural networks
dc.subjectNormative modeling
dc.subjectApplied mathematics
dc.subjectComputer science
dc.subjectNeurosciences
dc.subject.otherApplied mathematics
dc.titleDeep learning frameworks for modeling how neural circuits learn
dc.typeThesis

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Liu_washington_0250E_26903.pdf
Size:
18.62 MB
Format:
Adobe Portable Document Format