Word Embeddings

1106 papers with code • 0 benchmarks • 52 datasets

Word embedding is the collective name for a set of language modeling and feature learning techniques in natural language processing (NLP) where words or phrases from the vocabulary are mapped to vectors of real numbers.

Techniques for learning word embeddings can include Word2Vec, GloVe, and other neural network-based approaches that train on an NLP task such as language modeling or document classification.

( Image credit: Dynamic Word Embedding for Evolving Semantic Discovery )

Latest papers with no code

WordDecipher: Enhancing Digital Workspace Communication with Explainable AI for Non-native English Speakers

no code yet • 10 Apr 2024

Non-native English speakers (NNES) face challenges in digital workspace communication (e. g., emails, Slack messages), often inadvertently translating expressions from their native languages, which can lead to awkward or incorrect usage.

Robust Concept Erasure Using Task Vectors

no code yet • 4 Apr 2024

Finally, we show that Diverse Inversion enables us to apply a TV edit only to a subset of the model weights, enhancing the erasure capabilities while better maintaining the core functionality of the model.

PejorativITy: Disambiguating Pejorative Epithets to Improve Misogyny Detection in Italian Tweets

no code yet • 3 Apr 2024

Disambiguating the meaning of such terms might help the detection of misogyny.

The Shape of Word Embeddings: Recognizing Language Phylogenies through Topological Data Analysis

no code yet • 30 Mar 2024

Word embeddings represent language vocabularies as clouds of $d$-dimensional points.

Natural Language, AI, and Quantum Computing in 2024: Research Ingredients and Directions in QNLP

no code yet • 28 Mar 2024

Language processing is at the heart of current developments in artificial intelligence, and quantum computers are becoming available at the same time.

Fusion approaches for emotion recognition from speech using acoustic and text-based features

no code yet • 27 Mar 2024

In this paper, we study different approaches for classifying emotions from speech using acoustic and text-based features.

A comparative analysis of embedding models for patent similarity

no code yet • 25 Mar 2024

First, it compares the performance of different kinds of patent-specific pretrained embedding models, namely static word embeddings (such as word2vec and doc2vec models) and contextual word embeddings (such as transformers based models), on the task of patent similarity calculation.

An efficient domain-independent approach for supervised keyphrase extraction and ranking

no code yet • 24 Mar 2024

We present a supervised learning approach for automatic extraction of keyphrases from single documents.

Empowering Segmentation Ability to Multi-modal Large Language Models

no code yet • 21 Mar 2024

Multi-modal large language models (MLLMs) can understand image-language prompts and demonstrate impressive reasoning ability.

Leveraging Linguistically Enhanced Embeddings for Open Information Extraction

no code yet • 20 Mar 2024

To bridge this gap, we are the first to leverage linguistic features with a Seq2Seq PLM for OIE.