BERT Goes Shopping: Comparing Distributional Models for Product Representations

12/17/2020
by   Federico Bianchi, et al.
1

Word embeddings (e.g., word2vec) have been applied successfully to eCommerce products through prod2vec. Inspired by the recent performance improvements on several NLP tasks brought by contextualized embeddings, we propose to transfer BERT-like architectures to eCommerce: our model – ProdBERT – is trained to generate representations of products through masked session modeling. Through extensive experiments over multiple shops, different tasks, and a range of design choices, we systematically compare the accuracy of ProdBERT and prod2vec embeddings: while ProdBERT is found to be superior to traditional methods in several scenarios, we highlight the importance of resources and hyperparameters in the best performing models. Finally, we conclude by providing guidelines for training embeddings under a variety of computational and data constraints.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/02/2019

How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT, ELMo, and GPT-2 Embeddings

Replacing static word embeddings with contextualized word representation...
research
06/02/2023

Word Embeddings for Banking Industry

Applications of Natural Language Processing (NLP) are plentiful, from se...
research
04/02/2021

Query2Prod2Vec Grounded Word Embeddings for eCommerce

We present Query2Prod2Vec, a model that grounds lexical representations ...
research
01/25/2019

Word Embeddings: A Survey

This work lists and describes the main recent strategies for building fi...
research
08/08/2021

Efficacy of BERT embeddings on predicting disaster from Twitter data

Social media like Twitter provide a common platform to share and communi...
research
06/21/2022

NorBERT: NetwOrk Representations through BERT for Network Analysis and Management

Deep neural network models have been very successfully applied to Natura...
research
05/24/2023

You Are What You Annotate: Towards Better Models through Annotator Representations

Annotator disagreement is ubiquitous in natural language processing (NLP...

Please sign up or login with your details

Forgot password? Click here to reset