Multilingual NER Transfer for Low-resource Languages

02/01/2019
by   Afshin Rahimi, et al.
0

In massively multilingual transfer NLP models over many source languages are applied to a low-resource target language. In contrast to most prior work, which use a single model or a small handful, we consider many such models, which raises the critical problem of poor transfer, particularly from distant languages. We propose two techniques for modulating the transfer: one based on unsupervised truth inference, and another using limited supervision in the target language. Evaluating on named entity recognition over 41 languages, we show that our techniques are much more effective than strong baselines, including standard ensembling, and our unsupervised method rivals oracle selection of the single best individual model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/10/2023

Enhancing Low Resource NER Using Assisting Language And Transfer Learning

Named Entity Recognition (NER) is a fundamental task in NLP that is used...
research
10/07/2020

Transfer Learning and Distant Supervision for Multilingual Transformer Models: A Study on African Languages

Multilingual transformer models like mBERT and XLM-RoBERTa have obtained...
research
11/04/2022

Intriguing Properties of Compression on Multilingual Models

Multilingual models are often particularly dependent on scaling to gener...
research
12/15/2019

A Comparison of Architectures and Pretraining Methods for Contextualized Multilingual Word Embeddings

The lack of annotated data in many languages is a well-known challenge w...
research
10/10/2019

Language Transfer for Early Warning of Epidemics from Social Media

Statements on social media can be analysed to identify individuals who a...
research
10/10/2017

The Galactic Dependencies Treebanks: Getting More Data by Synthesizing New Languages

We release Galactic Dependencies 1.0---a large set of synthetic language...
research
02/15/2023

Meeting the Needs of Low-Resource Languages: The Value of Automatic Alignments via Pretrained Models

Large multilingual models have inspired a new class of word alignment me...

Please sign up or login with your details

Forgot password? Click here to reset