SOTAVerified

Word Embeddings

Word embedding is the collective name for a set of language modeling and feature learning techniques in natural language processing (NLP) where words or phrases from the vocabulary are mapped to vectors of real numbers.

Techniques for learning word embeddings can include Word2Vec, GloVe, and other neural network-based approaches that train on an NLP task such as language modeling or document classification.

( Image credit: Dynamic Word Embedding for Evolving Semantic Discovery )

Papers

Showing 471480 of 4002 papers

TitleStatusHype
Automated Detection of Adverse Drug Reactions in the Biomedical Literature Using Convolutional Neural Networks and Biomedical Word Embeddings0
An Analysis of Embedding Layers and Similarity Scores using Siamese Neural Networks0
Automated Discovery of Mathematical Definitions in Text0
Automated Discovery of Mathematical Definitions in Text with Deep Neural Networks0
Automated essay scoring with string kernels and word embeddings0
An Analysis of Hierarchical Text Classification Using Word Embeddings0
Automated Image Captioning for Rapid Prototyping and Resource Constrained Environments0
Automated PII Extraction from Social Media for Raising Privacy Awareness: A Deep Transfer Learning Approach0
A Minimalist Approach to Shallow Discourse Parsing and Implicit Relation Recognition0
BERTMap: A BERT-based Ontology Alignment System0
Show:102550
← PrevPage 48 of 401Next →

No leaderboard results yet.