Learnability of Autoregressive Transformers

dc.contributor.advisorSteinert-Threlkeld, Shane
dc.contributor.authorHong, Jeongyeob
dc.date.accessioned2026-02-05T19:37:29Z
dc.date.available2026-02-05T19:37:29Z
dc.date.issued2026-02-05
dc.date.submitted2025
dc.descriptionThesis (Master's)--University of Washington, 2025
dc.description.abstractThis paper explores the learning mechanism of a decoder-only transformer through the lens of human concept learning. We investigated whether decoder-only Transformers experience the simplicity bias, a human tendency to favor simpler representations. To do so, we create a pipeline that generates every task that a decoder-only transformer can learn and express with a given input symbol, length, and depth. Our initial results show no sufficient evidence for simplicity bias occurring in the autoregressive models. We end our paper with a discussion of other factors that can explain the learnability of transformers, such as the computational cost of each operation.
dc.embargo.termsOpen Access
dc.format.mimetypeapplication/pdf
dc.identifier.otherHong_washington_0250O_29153.pdf
dc.identifier.urihttps://hdl.handle.net/1773/55251
dc.language.isoen_US
dc.rightsCC BY
dc.subjectLinguistics
dc.subjectCognitive psychology
dc.subjectComputer science
dc.subject.otherLinguistics
dc.titleLearnability of Autoregressive Transformers
dc.typeThesis

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Hong_washington_0250O_29153.pdf
Size:
686.16 KB
Format:
Adobe Portable Document Format

Collections