Multilingual Pre-Trained Transformers and Convolutional NN Classification Models for Technical Domain Identification

01/22/2021
by   Suman Dowlagar, et al.
0

In this paper, we present a transfer learning system to perform technical domain identification on multilingual text data. We have submitted two runs, one uses the transformer model BERT, and the other uses XLM-ROBERTa with the CNN model for text classification. These models allowed us to identify the domain of the given sentences for the ICON 2020 shared Task, TechDOfication: Technical Domain Identification. Our system ranked the best for the subtasks 1d, 1g for the given TechDOfication dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/01/2021

SJ_AJ@DravidianLangTech-EACL2021: Task-Adaptive Pre-Training of Multilingual BERT models for Offensive Language Identification

In this paper we present our submission for the EACL 2021-Shared Task on...
research
07/21/2021

Comparison of Czech Transformers on Text Classification Tasks

In this paper, we present our progress in pre-training monolingual Trans...
research
02/24/2021

Multichannel LSTM-CNN for Telugu Technical Domain Identification

With the instantaneous growth of text information, retrieving domain-ori...
research
09/17/2021

The futility of STILTs for the classification of lexical borrowings in Spanish

The first edition of the IberLEF 2021 shared task on automatic detection...
research
05/18/2023

mLongT5: A Multilingual and Efficient Text-To-Text Transformer for Longer Sequences

We present our work on developing a multilingual, efficient text-to-text...
research
10/05/2021

ur-iw-hnt at GermEval 2021: An Ensembling Strategy with Multiple BERT Models

This paper describes our approach (ur-iw-hnt) for the Shared Task of Ger...
research
01/31/2021

Short Text Clustering with Transformers

Recent techniques for the task of short text clustering often rely on wo...

Please sign up or login with your details

Forgot password? Click here to reset