Adversarial Cross-Modal Retrieval via Learning and Transferring Single-Modal Similarities

04/17/2019
by   Zhizhong Han, et al.
0

Cross-modal retrieval aims to retrieve relevant data across different modalities (e.g., texts vs. images). The common strategy is to apply element-wise constraints between manually labeled pair-wise items to guide the generators to learn the semantic relationships between the modalities, so that the similar items can be projected close to each other in the common representation subspace. However, such constraints often fail to preserve the semantic structure between unpaired but semantically similar items (e.g. the unpaired items with the same class label are more similar than items with different labels). To address the above problem, we propose a novel cross-modal similarity transferring (CMST) method to learn and preserve the semantic relationships between unpaired items in an unsupervised way. The key idea is to learn the quantitative similarities in single-modal representation subspace, and then transfer them to the common representation subspace to establish the semantic relationships between unpaired items across modalities. Experiments show that our method outperforms the state-of-the-art approaches both in the class-based and pair-based retrieval tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/15/2022

Efficient Cross-Modal Retrieval via Deep Binary Hashing and Quantization

Cross-modal retrieval aims to search for data with similar semantic mean...
research
06/01/2017

Cross-modal Common Representation Learning by Hybrid Transfer Network

DNN-based cross-modal retrieval is a research hotspot to retrieve across...
research
10/10/2018

Temporal Cross-Media Retrieval with Soft-Smoothing

Multimedia information have strong temporal correlations that shape the ...
research
08/20/2018

An Efficient Approach for Geo-Multimedia Cross-Modal Retrieval

Due to the rapid development of mobile Internet techniques, cloud comput...
research
01/26/2022

Discriminative Supervised Subspace Learning for Cross-modal Retrieval

Nowadays the measure between heterogeneous data is still an open problem...
research
04/15/2023

CoVLR: Coordinating Cross-Modal Consistency and Intra-Modal Structure for Vision-Language Retrieval

Current vision-language retrieval aims to perform cross-modal instance s...
research
09/03/2021

Information Symmetry Matters: A Modal-Alternating Propagation Network for Few-Shot Learning

Semantic information provides intra-class consistency and inter-class di...

Please sign up or login with your details

Forgot password? Click here to reset