On the Effect of Pre-training for Transformer in Different Modality on Offline Reinforcement Learning

11/17/2022
by   Shiro Takagi, et al.
0

We empirically investigate how pre-training on data of different modalities, such as language and vision, affects fine-tuning of Transformer-based models to Mujoco offline reinforcement learning tasks. Analysis of the internal representation reveals that the pre-trained Transformers acquire largely different representations before and after pre-training, but acquire less information of data in fine-tuning than the randomly initialized one. A closer look at the parameter changes of the pre-trained Transformers reveals that their parameters do not change that much and that the bad performance of the model pre-trained with image data could partially come from large gradients and gradient clipping. To study what information the Transformer pre-trained with language data utilizes, we fine-tune this model with no context provided, finding that the model learns efficiently even without context information. Subsequent follow-up analysis supports the hypothesis that pre-training with language data is likely to make the Transformer get context-like information and utilize it to solve the downstream task.

READ FULL TEXT

page 21

page 22

page 25

page 32

page 33

page 34

page 35

page 41

research
04/23/2020

UHH-LT LT2 at SemEval-2020 Task 12: Fine-Tuning of Pre-Trained Transformer Networks for Offensive Language Detection

Fine-tuning of pre-trained transformer networks such as BERT yield state...
research
08/20/2023

Large Transformers are Better EEG Learners

Pre-trained large transformer models have achieved remarkable performanc...
research
05/30/2022

Multi-Game Decision Transformers

A longstanding goal of the field of AI is a strategy for compiling diver...
research
03/22/2022

MetaMorph: Learning Universal Controllers with Transformers

Multiple domains like vision, natural language, and audio are witnessing...
research
03/28/2023

TabRet: Pre-training Transformer-based Tabular Models for Unseen Columns

We present TabRet, a pre-trainable Transformer-based model for tabular d...
research
01/25/2022

Do Transformers Encode a Foundational Ontology? Probing Abstract Classes in Natural Language

With the methodological support of probing (or diagnostic classification...
research
07/20/2023

PASTA: Pretrained Action-State Transformer Agents

Self-supervised learning has brought about a revolutionary paradigm shif...

Please sign up or login with your details

Forgot password? Click here to reset