Fine-Tuning Transformers: Vocabulary Transfer

12/29/2021
by   Igor Samenko, et al.
0

Transformers are responsible for the vast majority of recent advances in natural language processing. The majority of practical natural language processing applications of these models is typically enabled through transfer learning. This paper studies if corpus-specific tokenization used for fine-tuning improves the resulting performance of the model. Through a series of experiments, we demonstrate that such tokenization combined with the initialization and fine-tuning strategy for the vocabulary tokens speeds up the transfer and boosts the performance of the fine-tuned model. We call this aspect of transfer facilitation vocabulary transfer.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/04/2022

Vocabulary Transfer for Medical Texts

Vocabulary transfer is a transfer learning subtask in which language mod...
research
06/04/2022

Actuarial Applications of Natural Language Processing Using Transformers: Case Studies for Using Text Features in an Actuarial Context

This tutorial demonstrates workflows to incorporate text data into actua...
research
04/26/2023

Fine Tuning with Abnormal Examples

Given the prevalence of crowd sourced labor in creating Natural Language...
research
02/09/2021

Transfer Learning Approach for Arabic Offensive Language Detection System – BERT-Based Model

Developing a system to detect online offensive language is very importan...
research
05/08/2021

Enhancing Transformers with Gradient Boosted Decision Trees for NLI Fine-Tuning

Transfer learning has become the dominant paradigm for many natural lang...
research
03/18/2022

Three things everyone should know about Vision Transformers

After their initial success in natural language processing, transformer ...
research
06/09/2021

URLTran: Improving Phishing URL Detection Using Transformers

Browsers often include security features to detect phishing web pages. I...

Please sign up or login with your details

Forgot password? Click here to reset