Feature Engineering vs BERT on Twitter Data

10/28/2022
by   Ryiaadh Gani, et al.
0

In this paper, we compare the performances of traditional machine learning models using feature engineering and word vectors and the state-of-the-art language model BERT using word embeddings on three datasets. We also consider the time and cost efficiency of feature engineering compared to BERT. From our results we conclude that the use of the BERT model was only worth the time and cost trade-off for one of the three datasets we used for comparison, where the BERT model significantly outperformed any kind of traditional classifier that uses feature vectors, instead of embeddings. Using the BERT model for the other datasets only achieved an increase of 0.03 and 0.05 of accuracy and F1 score respectively, which could be argued makes its use not worth the time and cost of GPU.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/07/2020

Dartmouth CS at WNUT-2020 Task 2: Informative COVID-19 Tweet Classification Using BERT

We describe the systems developed for the WNUT-2020 shared task 2, ident...
research
03/15/2022

Imputing Out-of-Vocabulary Embeddings with LOVE Makes Language Models Robust with Little Cost

State-of-the-art NLP systems represent inputs with word embeddings, but ...
research
01/25/2021

A Hybrid Approach to Measure Semantic Relatedness in Biomedical Concepts

Objective: This work aimed to demonstrate the effectiveness of a hybrid ...
research
05/22/2020

Comparative Study of Machine Learning Models and BERT on SQuAD

This study aims to provide a comparative analysis of performance of cert...
research
12/30/2019

AutoDiscern: Rating the Quality of Online Health Information with Hierarchical Encoder Attention-based Neural Networks

Patients increasingly turn to search engines and online content before, ...
research
10/08/2017

Clickbait detection using word embeddings

Clickbait is a pejorative term describing web content that is aimed at g...
research
10/10/2022

Multi-CLS BERT: An Efficient Alternative to Traditional Ensembling

Ensembling BERT models often significantly improves accuracy, but at the...

Please sign up or login with your details

Forgot password? Click here to reset