Full-attention based Neural Architecture Search using Context Auto-regression

11/13/2021
by   Yuan Zhou, et al.
19

Self-attention architectures have emerged as a recent advancement for improving the performance of vision tasks. Manual determination of the architecture for self-attention networks relies on the experience of experts and cannot automatically adapt to various scenarios. Meanwhile, neural architecture search (NAS) has significantly advanced the automatic design of neural architectures. Thus, it is appropriate to consider using NAS methods to discover a better self-attention architecture automatically. However, it is challenging to directly use existing NAS methods to search attention networks because of the uniform cell-based search space and the lack of long-term content dependencies. To address this issue, we propose a full-attention based NAS method. More specifically, a stage-wise search space is constructed that allows various attention operations to be adopted for different layers of a network. To extract global features, a self-supervised search algorithm is proposed that uses context auto-regression to discover the full-attention architecture. To verify the efficacy of the proposed methods, we conducted extensive experiments on various learning tasks, including image classification, fine-grained image recognition, and zero-shot image retrieval. The empirical results show strong evidence that our method is capable of discovering high-performance, full-attention architectures while guaranteeing the required search efficiency.

READ FULL TEXT
research
03/08/2023

HyT-NAS: Hybrid Transformers Neural Architecture Search for Edge Devices

Vision Transformers have enabled recent attention-based Deep Learning (D...
research
02/21/2021

Contrastive Self-supervised Neural Architecture Search

This paper proposes a novel cell-based neural architecture search algori...
research
03/14/2023

Learning to Grow Artificial Hippocampi in Vision Transformers for Resilient Lifelong Learning

Lifelong learning without catastrophic forgetting (i.e., resiliency) pos...
research
09/20/2019

Understanding Architectures Learnt by Cell-based Neural Architecture Search

Neural architecture search (NAS) generates architectures automatically f...
research
05/31/2019

Efficient Forward Architecture Search

We propose a neural architecture search (NAS) algorithm, Petridish, to i...
research
03/23/2019

Auto-ReID: Searching for a Part-aware ConvNet for Person Re-Identification

Prevailing deep convolutional neural networks (CNNs) for person re-IDent...
research
07/15/2021

AutoBERT-Zero: Evolving BERT Backbone from Scratch

Transformer-based pre-trained language models like BERT and its variants...

Please sign up or login with your details

Forgot password? Click here to reset