What is: Cross-View Training?
Source | Semi-Supervised Sequence Modeling with Cross-View Training |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Cross View Training, or CVT, is a semi-supervised algorithm for training distributed word representations that makes use of unlabelled and labelled examples.
CVT adds auxiliary prediction modules to the model, a Bi-LSTM encoder, which are used when learning on unlabeled examples. A prediction module is usually a small neural network (e.g., a hidden layer followed by a softmax layer). Each one takes as input an intermediate representation produced by the model (e.g., the outputs of one of the LSTMs in a Bi-LSTM model). It outputs a distribution over labels .
Each is chosen such that it only uses a part of the input ; the particular choice can depend on the task and model architecture. The auxiliary prediction modules are only used during training; the test-time prediction come from the primary prediction module that produces .