Investigating the significance of adversarial attacks and their relation to interpretability for radar-based human activity recognition systems

FOS: Computer and information sciences Computer Science - Machine Learning Technology and Engineering Computer Vision and Pattern Recognition (cs.CV) Computer Science - Computer Vision and Pattern Recognition 02 engineering and technology Machine Learning (cs.LG) Multimedia (cs.MM) Radar data Deep convolutional neural networks Activity recognition Adversarial examples 0202 electrical engineering, electronic engineering, information engineering Neural network interpretability Computer Science - Multimedia
DOI: 10.1016/j.cviu.2020.103111 Publication Date: 2020-09-23T21:38:01Z
ABSTRACT
Given their substantial success in addressing a wide range of computer vision challenges, Convolutional Neural Networks (CNNs) are increasingly being used in smart home applications, with many of these applications relying on the automatic recognition of human activities. In this context, low-power radar devices have recently gained in popularity as recording sensors, given that the usage of these devices allows mitigating a number of privacy concerns, a key issue when making use of conventional video cameras. Another concern that is often cited when designing smart home applications is the resilience of these applications against cyberattacks. It is, for instance, well-known that the combination of images and CNNs is vulnerable against adversarial examples, mischievous data points that force machine learning models to generate wrong classifications during testing time. In this paper, we investigate the vulnerability of radar-based CNNs to adversarial attacks, and where these radar-based CNNs have been designed to recognize human gestures. Through experiments with four unique threat models, we show that radar-based CNNs are susceptible to both white- and black-box adversarial attacks. We also expose the existence of an extreme adversarial attack case, where it is possible to change the prediction made by the radar-based CNNs by only perturbing the padding of the inputs, without touching the frames where the action itself occurs. Moreover, we observe that gradient-based attacks exercise perturbation not randomly, but on important features of the input data. We highlight these important features by making use of Grad-CAM, a popular neural network interpretability method, hereby showing the connection between adversarial perturbation and prediction interpretability.<br/>Accepted for publication on Computer Vision and Image Understanding, Special issue on Adversarial Deep Learning in Biometrics & Forensics<br/>
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (56)
CITATIONS (14)