What is: fastText?
Source | Enriching Word Vectors with Subword Information |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
fastText embeddings exploit subword information to construct word embeddings. Representations are learnt of character -grams, and words represented as the sum of the -gram vectors. This extends the word2vec type models with subword information. This helps the embeddings understand suffixes and prefixes. Once a word is represented using character -grams, a skipgram model is trained to learn the embeddings.