Knowledge-driven Natural Language Understanding
| dc.contributor.advisor | Xia, Fei | |
| dc.contributor.author | Tian, Yuanhe | |
| dc.date.accessioned | 2025-08-01T22:26:18Z | |
| dc.date.available | 2025-08-01T22:26:18Z | |
| dc.date.issued | 2025-08-01 | |
| dc.date.submitted | 2025 | |
| dc.description | Thesis (Ph.D.)--University of Washington, 2025 | |
| dc.description.abstract | Recent advances in natural language processing (NLP) mainly relied on pre-trained language models (PLMs) that are trained on vast amounts of data. Although these PLMs achieve remarkable success in improving NLP performance over conventional approaches, they still struggle to accurately understand the semantics of the text. Therefore, extra knowledge, especially the dynamically extracted one, is expected to be leveraged to improve the understanding of the models in the text. This thesis proposes a knowledge-driven approach for NLP that improves PLMs. By dynamically integrating external knowledge from multiple sources, the proposed approach enhances model generalization in different scenarios. Specifically, the thesis leverages three types of knowledge, namely, the lexicon knowledge (e.g., n-grams) extracted directly from raw data, syntax knowledge (e.g., dependency parse trees) obtained through existing toolkits, and the pattern knowledge (e.g., vectors) captured during the training process. Several novel architectures are proposed to leverage the knowledge, such as key-value memory networks for incorporating wordhood information, span attention mechanisms with categorical grouping for improved syntactic parsing, and graph convolutional networks to further enrich contextual representations. Extensive experiments on different NLP tasks at various levels demonstrate the effectiveness of the proposed approaches, which outperform strong baselines and existing studies. Overall, this dissertation not only broadens the definition and utilization of knowledge in natural language processing but also lays a solid foundation for future research in multi-modal, cross-domain, and low-resource environments. | |
| dc.embargo.terms | Open Access | |
| dc.format.mimetype | application/pdf | |
| dc.identifier.other | Tian_washington_0250E_28289.pdf | |
| dc.identifier.uri | https://hdl.handle.net/1773/53682 | |
| dc.language.iso | en_US | |
| dc.rights | CC BY | |
| dc.subject | Attention | |
| dc.subject | Knowledge | |
| dc.subject | Memory | |
| dc.subject | Natural Language Understanding | |
| dc.subject | Pre-trained Language Models | |
| dc.subject | Artificial intelligence | |
| dc.subject.other | Linguistics | |
| dc.title | Knowledge-driven Natural Language Understanding | |
| dc.type | Thesis |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Tian_washington_0250E_28289.pdf
- Size:
- 17.66 MB
- Format:
- Adobe Portable Document Format
