LSTAR Framework: Lightweight Framework for Standardizing Tests for Adversarial Robustness
Loading...
Date
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract
The role of neural networks in various tasks has exploded in recent years, becoming prevalentin many safety-critical applications. However, improving neural network robustness has be-
come a challenge due to the existence of adversarial examples—imperceptible perturbations
to the inputs of machine learning models that mislead classifiers into producing incorrect
outputs. While there have been numerous advancements in crafting adversarial attacks and
defenses, research on the basis of adversarial examples has notably lagged behind, largely
due to the computational difficulty of analyzing high-dimensional spaces. This inherent
difficulty has led researchers to construct models for understanding adversarial examples
divergent from conventional paradigms, with some relying on commonly used frameworks
while others utilize their own tailored frameworks to meet their unique needs. Consequently,
replicating and building upon research in this field presents a significant challenge. In this paper, we present a modular, lightweight framework to assist researchers in ad-dressing these challenges by providing a comprehensive approach to evaluating machine
learning models through a standardized experimentation platform. We present several po-
tential hypotheses regarding the basis of adversarial examples and utilize our framework to
verify them more robustly under complex attacks and datasets through controlled experi-
ments. Our experimental results indicate that geometric causes directly affect the robustness
of machine learning models, while statistical factors amplify the effects of adversarial attacks.
This framework provides a baseline for further studies to better understand the phenomenon
of adversarial examples, allowing researchers to design more robust machine learning models.
Description
Thesis (Master's)--University of Washington, 2024
