research
          
      
      ∙
      03/05/2020
    Distill, Adapt, Distill: Training Small, In-Domain Models for Neural Machine Translation
We explore best practices for training small, memory efficient machine t...
          
            research
          
      
      ∙
      02/19/2020
    Compressing BERT: Studying the Effects of Weight Pruning on Transfer Learning
Universal feature extractors, such as BERT for natural language processi...
          
            research
          
      
      ∙
      12/06/2019