Full metadata
Title
We Need to Talk About Robustness to Adversarial Attacks While Removing Spurious Dataset Biases
Description
Machine learning models can pick up biases and spurious correlations from training data and projects and amplify these biases during inference, thus posing significant challenges in real-world settings. One approach to mitigating this is a class of methods that can identify filter out bias-inducing samples from the training datasets to force models to avoid being exposed to biases. However, the filtering leads to a considerable wastage of resources as most of the dataset created is discarded as biased. This work deals with avoiding the wastage of resources by identifying and quantifying the biases. I further elaborate on the implications of dataset filtering on robustness (to adversarial attacks) and generalization (to out-of-distribution samples). The findings suggest that while dataset filtering does help to improve OOD(Out-Of-Distribution) generalization, it has a significant negative impact on robustness to adversarial attacks. It also shows that transforming bias-inducing samples into adversarial samples (instead of eliminating them from the dataset) can significantly boost robustness without sacrificing generalization.
Date Created
2021
Contributors
- Sachdeva, Bhavdeep Singh (Author)
- Baral, Chitta (Thesis advisor)
- Liu, Huan (Committee member)
- Yang, Yezhou (Committee member)
- Arizona State University (Publisher)
Topical Subject
Resource Type
Extent
72 pages
Language
eng
Copyright Statement
In Copyright
Primary Member of
Peer-reviewed
No
Open Access
No
Handle
https://hdl.handle.net/2286/R.2.N.161967
Level of coding
minimal
Cataloging Standards
Note
Partial requirement for: M.S., Arizona State University, 2021
Field of study: Computer Science
System Created
- 2021-11-16 05:34:14
System Modified
- 2021-11-30 12:51:28
- 2 years 11 months ago
Additional Formats