Distributional Semantics
Based on the idea that words appearing in similar contexts have similar meanings.
Skip-gram (in Word2Vec) – predicts context from a word
Learns word embeddings that reflect word meaning.
CBOW (Continuous Bag of Words) – predicts word from context
Similar goal, different direction.
GloVe (Global Vectors for Word Representation)
Combines matrix factorization with co-occurrence statistics.
Latent Semantic Analysis (LSA)
Reduces dimensionality of word-document matrices via SVD.
Latent Dirichlet Allocation (LDA)
A probabilistic model for discovering topics in text.
Contextual Embeddings (Neural-based semantics)
E.g., ELMo, BERT — generate embeddings based on the entire sentence.
Last updated