Improving the Adversarial Robustness of Transfer Learning via Noisy Feature Distillation

02/07/2020
by   Ting-Wu Chin, et al.
0

Fine-tuning through knowledge transfer from a pre-trained model on a large-scale dataset is a widely spread approach to effectively build models on small-scale datasets. However, recent literature has shown that such a fine-tuning approach is vulnerable to adversarial examples based on the pre-trained model, which raises security concerns for many industrial applications. In contrast, models trained with random initialization are much more robust to such attacks, although these models often exhibit much lower accuracy. In this work, we propose noisy feature distillation, a new transfer learning method that trains a network from random initialization while achieving clean-data performance competitive with fine-tuning. In addition, the method is shown empirically to significantly improve the robustness compared to fine-tuning with 15x reduction in attack success rate for ResNet-50, from 66 to 4.4 Actions, MIT 67 Indoor Scenes, and Oxford 102 Flowers datasets. Code is available at https://github.com/cmu-enyac/Renofeation.

READ FULL TEXT
research
03/20/2023

TWINS: A Fine-Tuning Framework for Improved Transferability of Adversarial Robustness and Generalization

Recent years have seen the ever-increasing importance of pre-trained mod...
research
05/24/2023

Refocusing Is Key to Transfer Learning

Transfer learning involves adapting a pre-trained model to novel downstr...
research
08/07/2019

Progressive Transfer Learning for Person Re-identification

Model fine-tuning is a widely used transfer learning approach in person ...
research
05/25/2019

Efficient Neural Task Adaptation by Maximum Entropy Initialization

Transferring knowledge from one neural network to another has been shown...
research
12/06/2016

Tag Prediction at Flickr: a View from the Darkroom

Automated photo tagging has established itself as one of the most compel...
research
11/11/2019

TANDA: Transfer and Adapt Pre-Trained Transformer Models for Answer Sentence Selection

We propose TANDA, an effective technique for fine-tuning pre-trained Tra...
research
07/14/2020

Automated Synthetic-to-Real Generalization

Models trained on synthetic images often face degraded generalization to...

Please sign up or login with your details

Forgot password? Click here to reset