FACM: Correct the Output of Deep Neural Network with Middle Layers Features against Adversarial Samples

06/02/2022
by   Xiangyuan Yang, et al.
0

In the strong adversarial attacks against deep neural network (DNN), the output of DNN will be misclassified if and only if the last feature layer of the DNN is completely destroyed by adversarial samples, while our studies found that the middle feature layers of the DNN can still extract the effective features of the original normal category in these adversarial attacks. To this end, in this paper, a middle Feature layer Analysis and Conditional Matching prediction distribution (FACM) model is proposed to increase the robustness of the DNN against adversarial samples through correcting the output of DNN with the features extracted by the middle layers of DNN. In particular, the middle Feature layer Analysis (FA) module, the conditional matching prediction distribution (CMPD) module and the output decision module are included in our FACM model to collaboratively correct the classification of adversarial samples. The experiments results show that, our FACM model can significantly improve the robustness of the naturally trained model against various attacks, and our FA model can significantly improve the robustness of the adversarially trained model against white-box attacks with weak transferability and black box attacks where FA model includes the FA module and the output decision module, not the CMPD module.

READ FULL TEXT

page 1

page 7

page 8

research
04/25/2023

Improving Robustness Against Adversarial Attacks with Deeply Quantized Neural Networks

Reducing the memory footprint of Machine Learning (ML) models, particula...
research
11/26/2020

Exposing the Robustness and Vulnerability of Hybrid 8T-6T SRAM Memory Architectures to Adversarial Attacks in Deep Neural Networks

Deep Learning is able to solve a plethora of once impossible problems. H...
research
12/08/2020

Mitigating the Impact of Adversarial Attacks in Very Deep Networks

Deep Neural Network (DNN) models have vulnerabilities related to securit...
research
05/21/2018

A Simple Cache Model for Image Recognition

Training large-scale image recognition models is computationally expensi...
research
11/19/2019

A novel method for identifying the deep neural network model with the Serial Number

Deep neural network (DNN) with the state of art performance has emerged ...
research
07/13/2019

Towards Characterizing and Limiting Information Exposure in DNN Layers

Pre-trained Deep Neural Network (DNN) models are increasingly used in sm...
research
04/14/2022

Q-TART: Quickly Training for Adversarial Robustness and in-Transferability

Raw deep neural network (DNN) performance is not enough; in real-world s...

Please sign up or login with your details

Forgot password? Click here to reset