Efficient Parallel Learning of Word2Vec

06/24/2016
by   Jeroen B. P. Vuurens, et al.
0

Since its introduction, Word2Vec and its variants are widely used to learn semantics-preserving representations of words or entities in an embedding space, which can be used to produce state-of-art results for various Natural Language Processing tasks. Existing implementations aim to learn efficiently by running multiple threads in parallel while operating on a single model in shared memory, ignoring incidental memory update collisions. We show that these collisions can degrade the efficiency of parallel learning, and propose a straightforward caching strategy that improves the efficiency by a factor of 4.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/09/2018

Learning Semantic Representations for Novel Words: Leveraging Both Form and Context

Word embeddings are a key component of high-performing natural language ...
research
08/09/2018

Deep Learning Based Natural Language Processing for End to End Speech Translation

Deep Learning methods employ multiple processing layers to learn hierarc...
research
07/25/2018

Shared-Memory Parallel Maximal Clique Enumeration

We present shared-memory parallel methods for Maximal Clique Enumeration...
research
11/07/2019

Probing Contextualized Sentence Representations with Visual Awareness

We present a universal framework to model contextualized sentence repres...
research
06/04/2019

Transferable Neural Projection Representations

Neural word representations are at the core of many state-of-the-art nat...
research
09/15/2021

ARCH: Efficient Adversarial Regularized Training with Caching

Adversarial regularization can improve model generalization in many natu...
research
08/18/2018

Maximising Throughput in a Complex Coal Export System

The Port of Newcastle features three coal export terminals, operating pr...

Please sign up or login with your details

Forgot password? Click here to reset