Lagesse, BrentTran, Kenneth2024-09-092024-09-092024-09-092024Tran_washington_0250O_26699.pdfhttps://hdl.handle.net/1773/51652Thesis (Master's)--University of Washington, 2024The role of neural networks in various tasks has exploded in recent years, becoming prevalentin many safety-critical applications. However, improving neural network robustness has be- come a challenge due to the existence of adversarial examples—imperceptible perturbations to the inputs of machine learning models that mislead classifiers into producing incorrect outputs. While there have been numerous advancements in crafting adversarial attacks and defenses, research on the basis of adversarial examples has notably lagged behind, largely due to the computational difficulty of analyzing high-dimensional spaces. This inherent difficulty has led researchers to construct models for understanding adversarial examples divergent from conventional paradigms, with some relying on commonly used frameworks while others utilize their own tailored frameworks to meet their unique needs. Consequently, replicating and building upon research in this field presents a significant challenge. In this paper, we present a modular, lightweight framework to assist researchers in ad-dressing these challenges by providing a comprehensive approach to evaluating machine learning models through a standardized experimentation platform. We present several po- tential hypotheses regarding the basis of adversarial examples and utilize our framework to verify them more robustly under complex attacks and datasets through controlled experi- ments. Our experimental results indicate that geometric causes directly affect the robustness of machine learning models, while statistical factors amplify the effects of adversarial attacks. This framework provides a baseline for further studies to better understand the phenomenon of adversarial examples, allowing researchers to design more robust machine learning models.application/pdfen-USCC BYAdversarial attacksAdversarial machine learningCuPyDeep learningNeural networksResearch standardizationComputer scienceComputing and software systemsLSTAR Framework: Lightweight Framework for Standardizing Tests for Adversarial RobustnessThesis