Choose an analogy:
Nearest Neighbor (Result)
queen
Cosine Similarity: 0.94
How Does It Work?
Embeddings learn semantic relationships during training. "king" and "queen" have similar vectors, but the difference to "man"/"woman" encodes gender. These difference vectors are consistent.
Vector Arithmetic
result = E["king"] - E["man"] + E["woman"]. We compute the nearest neighbor to this result vector (via cosine similarity), often it's the expected word.
Limitation
These analogies don't always work perfectly. They depend on training data quality and can reflect biases. Modern contextual embeddings (BERT, GPT) are more complex.
Historical Significance
Word2Vec (2013) made these analogies famous. They showed that neural embeddings learn semantic structure – a breakthrough in NLP.