File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Can Adversarial Weight Perturbations Inject Neural Backdoors

TitleCan Adversarial Weight Perturbations Inject Neural Backdoors
Authors
Keywordsadversarial deep learning
backdoor attacks
Issue Date2020
Citation
International Conference on Information and Knowledge Management, Proceedings, 2020, p. 2029-2032 How to Cite?
AbstractAdversarial machine learning has exposed several security hazards of neural models. Thus far, the concept of an "adversarial perturbation" has exclusively been used with reference to the input space referring to a small, imperceptible change which can cause a ML model to err. In this work we extend the idea of "adversarial perturbations" to the space of model weights, specifically to inject backdoors in trained DNNs, which exposes a security risk of publicly available trained models. Here, injecting a backdoor refers to obtaining a desired outcome from the model when a trigger pattern is added to the input, while retaining the original predictions on a non-triggered input. From the perspective of an adversary, we characterize these adversarial perturbations to be constrained within an ĝ.,"∞ norm around the original model weights. We introduce adversarial perturbations in model weights using a composite loss on the predictions of the original model and the desired trigger through projected gradient descent. Our results show that backdoors can be successfully injected with a very small average relative change in model weight values for several CV and NLP applications.
Persistent Identifierhttp://hdl.handle.net/10722/341291
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorGarg, Siddhant-
dc.contributor.authorKumar, Adarsh-
dc.contributor.authorGoel, Vibhor-
dc.contributor.authorLiang, Yingyu-
dc.date.accessioned2024-03-13T08:41:40Z-
dc.date.available2024-03-13T08:41:40Z-
dc.date.issued2020-
dc.identifier.citationInternational Conference on Information and Knowledge Management, Proceedings, 2020, p. 2029-2032-
dc.identifier.urihttp://hdl.handle.net/10722/341291-
dc.description.abstractAdversarial machine learning has exposed several security hazards of neural models. Thus far, the concept of an "adversarial perturbation" has exclusively been used with reference to the input space referring to a small, imperceptible change which can cause a ML model to err. In this work we extend the idea of "adversarial perturbations" to the space of model weights, specifically to inject backdoors in trained DNNs, which exposes a security risk of publicly available trained models. Here, injecting a backdoor refers to obtaining a desired outcome from the model when a trigger pattern is added to the input, while retaining the original predictions on a non-triggered input. From the perspective of an adversary, we characterize these adversarial perturbations to be constrained within an ĝ.,"∞ norm around the original model weights. We introduce adversarial perturbations in model weights using a composite loss on the predictions of the original model and the desired trigger through projected gradient descent. Our results show that backdoors can be successfully injected with a very small average relative change in model weight values for several CV and NLP applications.-
dc.languageeng-
dc.relation.ispartofInternational Conference on Information and Knowledge Management, Proceedings-
dc.subjectadversarial deep learning-
dc.subjectbackdoor attacks-
dc.titleCan Adversarial Weight Perturbations Inject Neural Backdoors-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1145/3340531.3412130-
dc.identifier.scopuseid_2-s2.0-85095864916-
dc.identifier.spage2029-
dc.identifier.epage2032-
dc.identifier.isiWOS:000749561302004-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats