Word Embeddings Revolution
Échec de l'ajout au panier.
Échec de l'ajout à la liste d'envies.
Échec de la suppression de la liste d’envies.
Échec du suivi du balado
Ne plus suivre le balado a échoué
-
Narrateur(s):
-
Auteur(s):
À propos de cet audio
In this episode, we explore the embedding revolution in natural language processing—the moment NLP moved from counting words to learning meaning. We trace how dense vector representations transformed language into a geometric space, enabling models to capture similarity, analogy, and semantic structure for the first time. This shift laid the groundwork for everything from modern search to large language models.
This episode covers:
• Why bag-of-words and TF-IDF failed to capture meaning
• The distributional hypothesis: “you know a word by the company it keeps”
• Dense vs. sparse representations and why geometry matters
• Topic models as early semantic compression (LSI, LDA)
• Word2Vec: CBOW and Skip-Gram
• Vector arithmetic and semantic analogies
• GloVe and global co-occurrence statistics
• FastText and subword representations
• The static ambiguity problem
• How embeddings led directly to RNNs, LSTMs, attention, and transformers
This episode is part of the Adapticx AI Podcast. Listen via the link provided or search “Adapticx” on Apple Podcasts, Spotify, Amazon Music, or most podcast platforms.
Sources and Further Reading
Additional references and extended material are available at: https://adapticx.co.uk