research
∙
03/05/2020
Distill, Adapt, Distill: Training Small, In-Domain Models for Neural Machine Translation
We explore best practices for training small, memory efficient machine t...
research
∙
02/19/2020
Compressing BERT: Studying the Effects of Weight Pruning on Transfer Learning
Universal feature extractors, such as BERT for natural language processi...
research
∙
12/06/2019