Word2vec
Word2vec is a word embedding model that learns the probability of some words being neighbours in a sentence $p_{neighbours}(w_i, w_o)$.
- Build a dataset of adjacent words. CBOW; skipgram; negative sampling;
- Encode the words using vectors.
- Build a model $f(\{\theta_i\})$ to calculate the probability of the words being neighours and improve the parameters $\{\theta_i\}$ using the dataset.
Planted:
by L Ma;
References:
Dynamic Backlinks to
wiki/machine-learning/embedding/word2vec
:wiki/machine-learning/embedding/word2vec
Links to:L Ma (2019). 'Word2vec', Datumorphism, 06 April. Available at: https://datumorphism.leima.is/wiki/machine-learning/embedding/word2vec/.