word2vec

Terms from Artificial Intelligence: humans at the heart of algorithms

Page numbers are for draft copy at present; they will be replaced with correct numbers when final book is formatted. Chapter numbers are correct and will not change now.

word2vec creates a vector representation of a word that captures certain aspects of the meaning of the word. It works by first creating a 'skip-gram model' that predicts the likelihood of surrounding words, and then extracts an internal layer of this model to form a latent space. The value of a word in this latent space is called a word vector. The geometry of the latent space often captures crucial relationships, so that, for example, the vector difference between 'dog' and 'puppy' is similar to that between 'cat' and 'kitten'. These word vectors can be used in further steps of an NLP process.

Used in Chap. 13: page 218; Chap. 21: page 361