Skip to main content

MODELGUARD: Information-Theoretic Defense Against Model Extraction Attacks

Publication ,  Conference
Tang, M; Dai, A; DiValentin, L; Ding, A; Hass, A; Gong, NZ; Chen, Y; Li, H
Published in: Proceedings of the 33rd USENIX Security Symposium
January 1, 2024

Malicious utilization of a query interface can compromise the confidentiality of ML-as-a-Service (MLaaS) systems via model extraction attacks. Previous studies have proposed to perturb the predictions of the MLaaS system as a defense against model extraction attacks. However, existing prediction perturbation methods suffer from a poor privacy-utility balance and cannot effectively defend against the latest adaptive model extraction attacks. In this paper, we propose a novel prediction perturbation defense named MODELGUARD, which aims at defending against adaptive model extraction attacks while maintaining a high utility of the protected system. We develop a general optimization problem that considers different kinds of model extraction attacks, and MODELGUARD provides an information-theoretic defense to efficiently solve the optimization problem and achieve resistance against adaptive attacks. Experiments show that MODELGUARD attains significantly better defensive performance against adaptive attacks with less loss of utility compared to previous defenses.

Duke Scholars

Published In

Proceedings of the 33rd USENIX Security Symposium

Publication Date

January 1, 2024

Start / End Page

5305 / 5322
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Tang, M., Dai, A., DiValentin, L., Ding, A., Hass, A., Gong, N. Z., … Li, H. (2024). MODELGUARD: Information-Theoretic Defense Against Model Extraction Attacks. In Proceedings of the 33rd USENIX Security Symposium (pp. 5305–5322).
Tang, M., A. Dai, L. DiValentin, A. Ding, A. Hass, N. Z. Gong, Y. Chen, and H. Li. “MODELGUARD: Information-Theoretic Defense Against Model Extraction Attacks.” In Proceedings of the 33rd USENIX Security Symposium, 5305–22, 2024.
Tang M, Dai A, DiValentin L, Ding A, Hass A, Gong NZ, et al. MODELGUARD: Information-Theoretic Defense Against Model Extraction Attacks. In: Proceedings of the 33rd USENIX Security Symposium. 2024. p. 5305–22.
Tang, M., et al. “MODELGUARD: Information-Theoretic Defense Against Model Extraction Attacks.” Proceedings of the 33rd USENIX Security Symposium, 2024, pp. 5305–22.
Tang M, Dai A, DiValentin L, Ding A, Hass A, Gong NZ, Chen Y, Li H. MODELGUARD: Information-Theoretic Defense Against Model Extraction Attacks. Proceedings of the 33rd USENIX Security Symposium. 2024. p. 5305–5322.

Published In

Proceedings of the 33rd USENIX Security Symposium

Publication Date

January 1, 2024

Start / End Page

5305 / 5322