Learning without Memorizing

11/20/2018
by   Prithviraj Dhar, et al.
0

Incremental learning (IL) is an important task aimed to increase the capability of a trained model, in terms of the number of classes recognizable by the model. The key problem in this task is the requirement of storing data (e.g. images) associated with existing classes, while training the classifier to learn new classes. However, this is impractical as it increases the memory requirement at every incremental step, which makes it impossible to implement IL algorithms on the edge devices with limited memory. Hence, we propose a novel approach, called "Learning without Memorizing (LwM)", to preserve the information with respect to existing (base) classes, without storing any of their data, while making the classifier progressively learn the new classes. In LwM, we present an information preserving penalty: Attention Distillation Loss, and demonstrate that penalizing the changes in classifiers' attention maps helps to retain information of the base classes, as new classes are added. We show that adding Attention Distillation Loss to the distillation loss which is an existing information preserving loss consistently outperforms the state-of-the-art performance in the iILSVRC-small and iCIFAR-100 datasets in terms of the overall accuracy of base and incrementally learned classes.

READ FULL TEXT

page 3

page 7

research
07/25/2018

End-to-End Incremental Learning

Although deep learning approaches have stood out in recent years due to ...
research
07/18/2022

Class-incremental Novel Class Discovery

We study the new task of class-incremental Novel Class Discovery (class-...
research
11/23/2016

iCaRL: Incremental Classifier and Representation Learning

A major open problem on the road to artificial intelligence is the devel...
research
01/16/2020

ScaIL: Classifier Weights Scaling for Class Incremental Learning

Incremental learning is useful if an AI agent needs to integrate data fr...
research
06/06/2022

Remember the Past: Distilling Datasets into Addressable Memories for Neural Networks

We propose an algorithm that compresses the critical information of a la...
research
04/26/2022

Improving Feature Generalizability with Multitask Learning in Class Incremental Learning

Many deep learning applications, like keyword spotting, require the inco...
research
10/19/2018

Generative Low-Shot Network Expansion

Conventional deep learning classifiers are static in the sense that they...

Please sign up or login with your details

Forgot password? Click here to reset