What is: TernaryBERT?
Source | TernaryBERT: Distillation-aware Ultra-low Bit BERT |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
TernaryBERT is a Transformer-based model which ternarizes the weights of a pretrained BERT model to , with different granularities for word embedding and weights in the Transformer layer. Instead of directly using knowledge distillation to compress a model, it is used to improve the performance of ternarized student model with the same size as the teacher model. In this way, we transfer the knowledge from the highly-accurate teacher model to the ternarized student model with smaller capacity.