Efficient Document Re-Ranking for Transformers by Precomputing Term Representations

04/29/2020
by   Sean MacAvaney, et al.
0

Deep pretrained transformer networks are effective at various ranking tasks, such as question answering and ad-hoc document ranking. However, their computational expenses deem them cost-prohibitive in practice. Our proposed approach, called PreTTR (Precomputing Transformer Term Representations), considerably reduces the query-time latency of deep transformer networks (up to a 42x speedup on web document ranking) making these networks more practical to use in a real-time ranking scenario. Specifically, we precompute part of the document term representations at indexing time without a query, and merge them with the query representation at query time to compute the final ranking score. Due to the large size of the token representations, we also propose an effective approach to reduce the storage requirement by training a compression layer to match attention scores. Our compression technique reduces the storage required up to 95 ranking performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/03/2019

TU Wien @ TREC Deep Learning '19 – Simple Contextualization for Re-ranking

The usage of neural network models puts multiple objectives in conflict ...
research
10/03/2021

SDR: Efficient Neural Re-ranking using Succinct Document Representation

BERT based ranking models have achieved superior performance on various ...
research
04/28/2020

EARL: Speedup Transformer-based Rankers with Pre-computed Representation

Recent innovations in Transformer-based ranking models have advanced the...
research
09/07/2015

Integrate Document Ranking Information into Confidence Measure Calculation for Spoken Term Detection

This paper proposes an algorithm to improve the calculation of confidenc...
research
03/29/2022

Compact Token Representations with Contextual Quantization for Efficient Document Re-ranking

Transformer based re-ranking models can achieve high search relevance th...
research
04/19/2021

Improving Transformer-Kernel Ranking Model Using Conformer and Query Term Independence

The Transformer-Kernel (TK) model has demonstrated strong reranking perf...
research
07/16/2018

Repeatability Corner Cases in Document Ranking: The Impact of Score Ties

Document ranking experiments should be repeatable: running the same rank...

Please sign up or login with your details

Forgot password? Click here to reset