Defending against attacks tailored to transfer learning via feature distancing
SCIE
SCOPUS
- Title
- Defending against attacks tailored to transfer learning via feature distancing
- Authors
- Ji, Sangwoo; Park, Namgyu; Na, Dongbin; Zhu, Bin; Kim, Jong
- Date Issued
- 2022-10
- Publisher
- Academic Press Inc.
- Abstract
- © 2022 Elsevier Inc.Transfer learning is preferable for training a deep neural network with a small training dataset by leveraging a pre-trained teacher model. However, transfer learning opens a door for new attacks that generate adversarial examples using the pre-trained teacher model. In this paper, we propose a novel method called feature distancing to defend against adversarial attacks tailored to transfer learning. The method aims to train a student model with a distinct feature representation from the teacher model. We generate adversarial examples of the mimic attack with the teacher model, and the examples are used to train the student model. We use triplet loss to put the mimic attack examples close to their source images and far from their target images in the feature space of the student model. The proposed method is evaluated on three different transfer learning tasks with diverse attack configurations. It is the only method that achieves high “robust accuracy” and high “test accuracy” on every task we evaluate.
- URI
- https://oasis.postech.ac.kr/handle/2014.oak/117883
- DOI
- 10.1016/j.cviu.2022.103533
- ISSN
- 1077-3142
- Article Type
- Article
- Citation
- Computer Vision and Image Understanding, vol. 223, 2022-10
- Files in This Item:
- There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.