Class-Specific Attention (CSA) for Time-Series Classification

11/19/2022
by   Yifan Hao, et al.
0

Most neural network-based classifiers extract features using several hidden layers and make predictions at the output layer by utilizing these extracted features. We observe that not all features are equally pronounced in all classes; we call such features class-specific features. Existing models do not fully utilize the class-specific differences in features as they feed all extracted features from the hidden layers equally to the output layers. Recent attention mechanisms allow giving different emphasis (or attention) to different features, but these attention models are themselves class-agnostic. In this paper, we propose a novel class-specific attention (CSA) module to capture significant class-specific features and improve the overall classification performance of time series. The CSA module is designed in a way such that it can be adopted in existing neural network (NN) based models to conduct time series classification. In the experiments, this module is plugged into five start-of-the-art neural network models for time series classification to test its effectiveness by using 40 different real datasets. Extensive experiments show that an NN model embedded with the CSA module can improve the base model in most cases and the accuracy improvement can be up to 42 statistical analysis show that the performance of an NN model embedding the CSA module is better than the base NN model on 67 and is significantly better on 11

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/02/2021

Learnable Dynamic Temporal Pooling for Time Series Classification

With the increase of available time series data, predicting their class ...
research
05/26/2020

A Protection against the Extraction of Neural Network Models

Given oracle access to a Neural Network (NN), it is possible to extract ...
research
03/18/2021

Linear Iterative Feature Embedding: An Ensemble Framework for Interpretable Model

A new ensemble framework for interpretable model called Linear Iterative...
research
04/22/2022

Exploring Hidden Semantics in Neural Networks with Symbolic Regression

Many recent studies focus on developing mechanisms to explain the black-...
research
07/14/2022

Rethinking Attention Mechanism in Time Series Classification

Attention-based models have been widely used in many areas, such as comp...
research
12/22/2020

Residual Matrix Product State for Machine Learning

Tensor network (TN), which originates from quantum physics, shows broad ...

Please sign up or login with your details

Forgot password? Click here to reset