HAT-CL: A Hard-Attention-to-the-Task PyTorch Library for Continual Learning

07/18/2023
by   Xiaotian Duan, et al.
0

Catastrophic forgetting, the phenomenon in which a neural network loses previously obtained knowledge during the learning of new tasks, poses a significant challenge in continual learning. The Hard-Attention-to-the-Task (HAT) mechanism has shown potential in mitigating this problem, but its practical implementation has been complicated by issues of usability and compatibility, and a lack of support for existing network reuse. In this paper, we introduce HAT-CL, a user-friendly, PyTorch-compatible redesign of the HAT mechanism. HAT-CL not only automates gradient manipulation but also streamlines the transformation of PyTorch modules into HAT modules. It achieves this by providing a comprehensive suite of modules that can be seamlessly integrated into existing architectures. Additionally, HAT-CL offers ready-to-use HAT networks that are smoothly integrated with the TIMM library. Beyond the redesign and reimplementation of HAT, we also introduce novel mask manipulation techniques for HAT, which have consistently shown improvements across various experiments. Our work paves the way for a broader application of the HAT mechanism, opening up new possibilities in continual learning across diverse models and applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/10/2019

Bayesian Optimized Continual Learning with Attention Mechanism

Though neural networks have achieved much progress in various applicatio...
research
08/27/2023

Universal Graph Continual Learning

We address catastrophic forgetting issues in graph learning as incoming ...
research
08/10/2022

ATLAS: Universal Function Approximator for Memory Retention

Artificial neural networks (ANNs), despite their universal function appr...
research
03/20/2022

Continual Sequence Generation with Adaptive Compositional Modules

Continual learning is essential for real-world deployment when there is ...
research
03/31/2020

Conditional Channel Gated Networks for Task-Aware Continual Learning

Convolutional Neural Networks experience catastrophic forgetting when op...
research
07/16/2023

A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning

Forgetting refers to the loss or deterioration of previously acquired in...
research
08/26/2021

Continual learning under domain transfer with sparse synaptic bursting

Existing machines are functionally specific tools that were made for eas...

Please sign up or login with your details

Forgot password? Click here to reset