dc.contributor.advisor | Steinert-Threlkeld, Shane | |
dc.contributor.author | Simpson, Katya | |
dc.date.accessioned | 2022-04-19T23:46:25Z | |
dc.date.available | 2022-04-19T23:46:25Z | |
dc.date.submitted | 2022 | |
dc.identifier.other | Simpson_washington_0250O_23979.pdf | |
dc.identifier.uri | http://hdl.handle.net/1773/48525 | |
dc.description | Thesis (Master's)--University of Washington, 2022 | |
dc.description.abstract | This thesis examines topical consistency between claims and fact-checks in the Birdwatch dataset published by Twitter. The dataset has tweets (the claims), notes (context-adding annotations written by Birdwatch users), and quality labels (ratings from the community of Birdwatch users). High quality notes can be thought of as potential “fact-checks” on the tweets. We find topics by clustering contextual word type embeddings (following a method introduced by Sia et al. [2020]) and evaluate two research questions: (1) Do notes that have high topic overlap with their associated tweet get better ratings? and (2) Can this topic modeling method be used to measure the helpful extra context that Birdwatch notes add to a tweet? Kullback-Leibler divergence is used to proxy topic overlap between the documents. We find that there is a statistically significant relationship between topic overlap and helpfulness but cannot establish a relationship between helpfulness and added context. | |
dc.format.mimetype | application/pdf | |
dc.language.iso | en_US | |
dc.rights | CC BY | |
dc.subject | Clustering | |
dc.subject | Embeddings | |
dc.subject | Fact-Checking | |
dc.subject | Misinformation | |
dc.subject | NLP | |
dc.subject | Topic Models | |
dc.subject | Linguistics | |
dc.subject | Information technology | |
dc.subject | Computer science | |
dc.subject.other | Linguistics | |
dc.title | "Obama never said that": Evaluating fact-checks for topical consistency and quality | |
dc.type | Thesis | |
dc.embargo.terms | Open Access | |