An Analysis of Word2Vec for the Italian Language

01/25/2020
by   Giovanni Di Gennaro, et al.
0

Word representation is fundamental in NLP tasks, because it is precisely from the coding of semantic closeness between words that it is possible to think of teaching a machine to understand text. Despite the spread of word embedding concepts, still few are the achievements in linguistic contexts other than English. In this work, analysing the semantic capacity of the Word2Vec algorithm, an embedding for the Italian language is produced. Parameter setting such as the number of epochs, the size of the context window and the number of negatively backpropagated samples is explored.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/23/2020

Musical Word Embedding: Bridging the Gap between Listening Contexts and Music

Word embedding pioneered by Mikolov et al. is a staple technique for wor...
research
09/15/2021

Fast Extraction of Word Embedding from Q-contexts

The notion of word embedding plays a fundamental role in natural languag...
research
12/11/2018

Delta Embedding Learning

Learning from corpus and learning from supervised NLP tasks both give us...
research
04/01/2019

Syntactic Interchangeability in Word Embedding Models

Nearest neighbors in word embedding models are commonly observed to be s...
research
02/12/2017

Vector Embedding of Wikipedia Concepts and Entities

Using deep learning for different machine learning tasks such as image c...
research
12/05/2019

Information Privacy Opinions on Twitter: A Cross-Language Study

The Cambridge Analytica scandal triggered a conversation on Twitter abou...
research
02/20/2020

FrameAxis: Characterizing Framing Bias and Intensity with Word Embedding

We propose FrameAxis, a method of characterizing the framing of a given ...

Please sign up or login with your details

Forgot password? Click here to reset