What is: MPNet?
Source | MPNet: Masked and Permuted Pre-training for Language Understanding |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
MPNet is a pre-training method for language models that combines masked language modeling (MLM) and permuted language modeling (PLM) in one view. It takes the dependency among the predicted tokens into consideration through permuted language modeling and thus avoids the issue of BERT. On the other hand, it takes position information of all tokens as input to make the model see the position information of all the tokens and thus alleviates the position discrepancy of XLNet.
The training objective of MPNet is:
As can be seen, MPNet conditions on (the tokens preceding the current predicted token ) rather than only the non-predicted tokens in MLM; comparing with PLM, MPNet takes more information (i.e., the mask symbol in position ) as inputs. Although the objective seems simple, it is challenging to implement the model efficiently. For details, see the paper.