Preserved Structure Across Vector Space Representations

02/02/2018
by   Andrei Amatuni, et al.
0

Certain concepts, words, and images are intuitively more similar than others (dog vs. cat, dog vs. spoon), though quantifying such similarity is notoriously difficult. Indeed, this kind of computation is likely a critical part of learning the category boundaries for words within a given language. Here, we use a set of 27 items (e.g. 'dog') that are highly common in infants' input, and use both image- and word-based algorithms to independently compute similarity among them. We find three key results. First, the pairwise item similarities derived within image-space and word-space are correlated, suggesting preserved structure among these extremely different representational formats. Second, the closest 'neighbors' for each item, within each space, showed significant overlap (e.g. both found 'egg' as a neighbor of 'apple'). Third, items with the most overlapping neighbors are later-learned by infants and toddlers. We conclude that this approach, which does not rely on human ratings of similarity, may nevertheless reflect stable within-class structure across these two spaces. We speculate that such invariance might aid lexical acquisition, by serving as an informative marker of category boundaries.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/06/2020

Embedding Words in Non-Vector Space with Unsupervised Graph Learning

It has become a de-facto standard to represent words as elements of a ve...
research
01/20/2016

Semantic Word Clusters Using Signed Normalized Graph Cuts

Vector space representations of words capture many aspects of word simil...
research
03/10/2023

Semi-supervised Adversarial Learning for Complementary Item Recommendation

Complementary item recommendations are a ubiquitous feature of modern e-...
research
12/07/2018

Gated Attentive-Autoencoder for Content-Aware Recommendation

The rapid growth of Internet services and mobile devices provides an exc...
research
12/03/2019

Modelling Semantic Categories using Conceptual Neighborhood

While many methods for learning vector space embeddings have been propos...
research
12/28/2021

Simple, Interpretable and Stable Method for Detecting Words with Usage Change across Corpora

The problem of comparing two bodies of text and searching for words that...
research
10/28/2022

You can't pick your neighbors, or can you? When and how to rely on retrieval in the kNN-LM

Retrieval-enhanced language models (LMs), which condition their predicti...

Please sign up or login with your details

Forgot password? Click here to reset