Skip to main content

Adversarial Training for Raw-Binary Malware Classifiers

Publication ,  Conference
Lucas, K; Bauer, L; Pai, S; Reiter, MK; Lin, W; Sharif, M
Published in: 32nd USENIX Security Symposium, USENIX Security 2023
January 1, 2023

Machine learning (ML) models have shown promise in classifying raw executable files (binaries) as malicious or benign with high accuracy. This has led to the increasing influence of ML-based classification methods in academic and real-world malware detection, a critical tool in cybersecurity. However, previous work provoked caution by creating variants of malicious binaries, referred to as adversarial examples, that are transformed in a functionality-preserving way to evade detection. In this work, we investigate the effectiveness of using adversarial training methods to create malware-classification models that are more robust to some state-of-the-art attacks. To train our most robust models, we significantly increase the efficiency and scale of creating adversarial examples to make adversarial training practical, which has not been done before in raw-binary malware detectors. We then analyze the effects of varying the length of adversarial training, as well as analyze the effects of training with various types of attacks. We find that data augmentation does not deter state-of-the-art attacks, but that using a generic gradient-guided method, used in other discrete domains, does improve robustness. We also show that in most cases, models can be made more robust to malware-domain attacks by adversarially training them with lower-effort versions of the same attack. In the best case, we reduce one state-of-the-art attack's success rate from 90% to 5%. We also find that training with some types of attacks can increase robustness to other types of attacks. Finally, we discuss insights gained from our results, and how they can be used to more effectively train robust malware detectors.

Duke Scholars

Published In

32nd USENIX Security Symposium, USENIX Security 2023

Publication Date

January 1, 2023

Volume

2

Start / End Page

1163 / 1180
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Lucas, K., Bauer, L., Pai, S., Reiter, M. K., Lin, W., & Sharif, M. (2023). Adversarial Training for Raw-Binary Malware Classifiers. In 32nd USENIX Security Symposium, USENIX Security 2023 (Vol. 2, pp. 1163–1180).
Lucas, K., L. Bauer, S. Pai, M. K. Reiter, W. Lin, and M. Sharif. “Adversarial Training for Raw-Binary Malware Classifiers.” In 32nd USENIX Security Symposium, USENIX Security 2023, 2:1163–80, 2023.
Lucas K, Bauer L, Pai S, Reiter MK, Lin W, Sharif M. Adversarial Training for Raw-Binary Malware Classifiers. In: 32nd USENIX Security Symposium, USENIX Security 2023. 2023. p. 1163–80.
Lucas, K., et al. “Adversarial Training for Raw-Binary Malware Classifiers.” 32nd USENIX Security Symposium, USENIX Security 2023, vol. 2, 2023, pp. 1163–80.
Lucas K, Bauer L, Pai S, Reiter MK, Lin W, Sharif M. Adversarial Training for Raw-Binary Malware Classifiers. 32nd USENIX Security Symposium, USENIX Security 2023. 2023. p. 1163–1180.

Published In

32nd USENIX Security Symposium, USENIX Security 2023

Publication Date

January 1, 2023

Volume

2

Start / End Page

1163 / 1180