Logo
Pattern

Discover published sets by community

Explore tens of thousands of sets crafted by our community.

Word Embedding Techniques

10

Flashcards

0/10

Still learning
StarStarStarStar

GloVe

StarStarStarStar

Global Vectors for Word Representation. Incorporates global statistics of the corpus by performing dimensionality reduction on the co-occurrence matrix of words to capture semantics.

StarStarStarStar

BERT

StarStarStarStar

Bidirectional Encoder Representations from Transformers. Generates deep bidirectional representations by jointly conditioning on both left and right context in all layers, primarily designed for fine-tuning on downstream tasks.

StarStarStarStar

ELMo

StarStarStarStar

Embeddings from Language Models. Utilizes a deep, bidirectional LSTM network to create contextualized word embeddings by taking into account the entire context in which a word appears.

StarStarStarStar

Transformer

StarStarStarStar

A model architecture that relies solely on attention mechanisms without recurrence, designed for parallelization and speed which has become the basis for many state-of-the-art NLP models.

StarStarStarStar

TF-IDF

StarStarStarStar

Term Frequency-Inverse Document Frequency. Weighs the frequency of a word in a document against its frequency across many documents to identify importance. Often used in document retrieval and indexing.

StarStarStarStar

GPT

StarStarStarStar

Generative Pretrained Transformer. Utilizes unsupervised learning for pretraining a transformer network on a large corpus, and is then fine-tuned for various tasks with supervised learning.

StarStarStarStar

One-Hot Encoding

StarStarStarStar

Represents words as sparse binary vectors where each word is represented by a 1 in its position and 0s elsewhere. Not used for capturing context or semantic information.

StarStarStarStar

FastText

StarStarStarStar

An extension of Word2Vec that treats each word as composed of character n-grams, which helps in capturing morphological information, especially for languages with rich word forms.

StarStarStarStar

Word2Vec

StarStarStarStar

A neural network based technique that learns distributed representations of words that capture semantic and syntactic patterns, through either the CBOW or Skip-gram models.

StarStarStarStar

CoVe

StarStarStarStar

Context Vectors. Uses a sequence-to-sequence model to provide contextually informed word representations by encoding the semantic meaning of words in context.

Know
0
Still learning
Click to flip
Know
0
Logo

© Hypatia.Tech. 2024 All rights reserved.