Crosslingual Sharing for Low-Resource Natural Language Processing

dc.contributor.advisorSmith, Noah A.
dc.contributor.authorMulcaire, Phoebe
dc.date.accessioned2022-07-14T22:08:08Z
dc.date.available2022-07-14T22:08:08Z
dc.date.issued2022-07-14
dc.date.submitted2022
dc.descriptionThesis (Ph.D.)--University of Washington, 2022
dc.description.abstractModern NLP systems have been highly successful at a wide variety of tasks, including language modeling and structured prediction problems such as syntactic and semantic parsing. This is due in large part to the use of supervised neural networks and more recently to unsupervised contextualized representations. However, these techniques rely on resources, such as extensive task-specific annotation and vast amounts of unlabeled text, which are not available in every language. Thus, most prior research has been focused on high-resource languages such as English. Crosslingual transfer from a high-resource source language, or sharing among many languages, has increasingly been used to achieve similar improvements in low-resource languages by exploiting underlying similarities between languages; however, many techniques for crosslingual transfer in turn require crosslingual resources such as parallel corpora, which again may not be available. All of these factors pose challenges to natural language processing in low-resource languages. This thesis argues that even with little, indirect or absent crosslingual supervision, sharing information between languages is a highly effective strategy for low-resource NLP, and quantifies the benefits in various low-resource settings and languages. We describe two lines of work addressing the problem of crosslingual transfer in such low-resource settings. In the first, we present language models and supervised structured prediction models which take a joint training approach, sharing parameters across several languages, to improve performance relative to monolingual training. We begin the second with GroC, a language model with compositional input and output representations which store linguistic information independently of any specific vocabulary, and show that GroC succeeds in low-resource language modeling and monolingual domain adaptation. Finally, we unite these two threads by using joint crosslingual training for compositional language models, including ones which use crosslingual lexicons not available to previous multilingual models. We show that this combined approach improves low-resource learning for a variety of target languages.
dc.embargo.termsOpen Access
dc.format.mimetypeapplication/pdf
dc.identifier.otherMulcaire_washington_0250E_24163.pdf
dc.identifier.urihttp://hdl.handle.net/1773/48884
dc.language.isoen_US
dc.relation.haspartCertificateOfCompletion.pdf; pdf; .
dc.rightsCC BY-NC-SA
dc.subject
dc.subjectComputer science
dc.subject.otherComputer science and engineering
dc.titleCrosslingual Sharing for Low-Resource Natural Language Processing
dc.typeThesis

Files

Original bundle

Now showing 1 - 2 of 2
Loading...
Thumbnail Image
Name:
Mulcaire_washington_0250E_24163.pdf
Size:
2.09 MB
Format:
Adobe Portable Document Format
Loading...
Thumbnail Image
Name:
CertificateOfCompletion.pdf
Size:
120.96 KB
Format:
Adobe Portable Document Format