Robust image classification against adversarial attacks using elastic similarity measures between edge count sequences

Copyright © 2020 Elsevier Ltd. All rights reserved..

Due to their unprecedented capacity to learn patterns from raw data, deep neural networks have become the de facto modeling choice to address complex machine learning tasks. However, recent works have emphasized the vulnerability of deep neural networks when being fed with intelligently manipulated adversarial data instances tailored to confuse the model. In order to overcome this issue, a major effort has been made to find methods capable of making deep learning models robust against adversarial inputs. This work presents a new perspective for improving the robustness of deep neural networks in image classification. In computer vision scenarios, adversarial images are crafted by manipulating legitimate inputs so that the target classifier is eventually fooled, but the manipulation is not visually distinguishable by an external observer. The reason for the imperceptibility of the attack is that the human visual system fails to detect minor variations in color space, but excels at detecting anomalies in geometric shapes. We capitalize on this fact by extracting color gradient features from input images at multiple sensitivity levels to detect possible manipulations. We resort to a deep neural classifier to predict the category of unseen images, whereas a discrimination model analyzes the extracted color gradient features with time series techniques to determine the legitimacy of input images. The performance of our method is assessed over experiments comprising state-of-the-art techniques for crafting adversarial attacks. Results corroborate the increased robustness of the classifier when using our discrimination module, yielding drastically reduced success rates of adversarial attacks that operate on the whole image rather than on localized regions or around the existing shapes of the image. Future research is outlined towards improving the detection accuracy of the proposed method for more general attack strategies.

Medienart:

E-Artikel

Erscheinungsjahr:

2020

Erschienen:

2020

Enthalten in:

Zur Gesamtaufnahme - volume:128

Enthalten in:

Neural networks : the official journal of the International Neural Network Society - 128(2020) vom: 01. Aug., Seite 61-72

Sprache:

Englisch

Beteiligte Personen:

Oregi, Izaskun [VerfasserIn]
Del Ser, Javier [VerfasserIn]
Pérez, Aritz [VerfasserIn]
Lozano, José A [VerfasserIn]

Links:

Volltext

Themen:

Adversarial machine learning
Computer vision
Deep neural networks
Journal Article
Time series analysis

Anmerkungen:

Date Completed 26.10.2020

Date Revised 26.10.2020

published: Print-Electronic

Citation Status MEDLINE

doi:

10.1016/j.neunet.2020.04.030

funding:

Förderinstitution / Projekttitel:

PPN (Katalog-ID):

NLM310248663