TY - JOUR
T1 - PICK-OBJECT-ATTACK
T2 - type-specific adversarial attack for object detection
AU - Mohamad Nezami, Omid
AU - Chaturvedi, Akshay
AU - Dras, Mark
AU - Garain, Utpal
PY - 2021/10
Y1 - 2021/10
N2 - Many recent studies have shown that deep neural models are vulnerable to adversarial samples: images with imperceptible perturbations, for example, can fool image classifiers. In this paper, we present the first type-specific approach to generating adversarial examples for object detection, which entails detecting bounding boxes around multiple objects present in the image and classifying them at the same time, making it a harder task than against image classification. We specifically aim to attack the widely used Faster R-CNN by changing the predicted label for a particular object in an image: where prior work has targeted one specific object (a stop sign), we generalize to arbitrary objects, with the key challenge being the need to change the labels of all bounding boxes for all instances of that object type. To do so, we propose a novel method, named PICK-OBJECT-ATTACK. PICK-OBJECT-ATTACK successfully adds perturbations only to bounding boxes for the targeted object, preserving the labels of other detected objects in the image. In terms of perceptibility, the perturbations induced by the method are very small. Furthermore, for the first time, we examine the effect of adversarial attacks on object detection in terms of a downstream task, image captioning; we show that where a method that can modify all object types leads to very obvious changes in captions, the changes from our constrained attack are much less apparent.
AB - Many recent studies have shown that deep neural models are vulnerable to adversarial samples: images with imperceptible perturbations, for example, can fool image classifiers. In this paper, we present the first type-specific approach to generating adversarial examples for object detection, which entails detecting bounding boxes around multiple objects present in the image and classifying them at the same time, making it a harder task than against image classification. We specifically aim to attack the widely used Faster R-CNN by changing the predicted label for a particular object in an image: where prior work has targeted one specific object (a stop sign), we generalize to arbitrary objects, with the key challenge being the need to change the labels of all bounding boxes for all instances of that object type. To do so, we propose a novel method, named PICK-OBJECT-ATTACK. PICK-OBJECT-ATTACK successfully adds perturbations only to bounding boxes for the targeted object, preserving the labels of other detected objects in the image. In terms of perceptibility, the perturbations induced by the method are very small. Furthermore, for the first time, we examine the effect of adversarial attacks on object detection in terms of a downstream task, image captioning; we show that where a method that can modify all object types leads to very obvious changes in captions, the changes from our constrained attack are much less apparent.
KW - Adversarial attack
KW - Faster R-CNN
KW - Deep learning
KW - Image captioning
KW - Computer vision
UR - http://www.scopus.com/inward/record.url?scp=85113699794&partnerID=8YFLogxK
U2 - 10.1016/j.cviu.2021.103257
DO - 10.1016/j.cviu.2021.103257
M3 - Article
AN - SCOPUS:85113699794
SN - 1077-3142
VL - 211
SP - 1
EP - 7
JO - Computer Vision and Image Understanding
JF - Computer Vision and Image Understanding
M1 - 103257
ER -