6th Place Solution to Google Universal Image Embedding

10/17/2022
by   S. Gkelios, et al.
0

This paper presents the 6th place solution to the Google Universal Image Embedding competition on Kaggle. Our approach is based on the CLIP architecture, a powerful pre-trained model used to learn visual representation from natural language supervision. We also utilized the SubCenter ArcFace loss with dynamic margins to improve the distinctive power of class separability and embeddings. Finally, a diverse dataset has been created based on the test's set categories and the leaderboard's feedback. By carefully crafting a training scheme to enhance transfer learning, our submission scored 0.685 on the private leaderboard.

READ FULL TEXT

page 1

page 2

page 3

research
10/14/2022

3rd Place Solution for Google Universal Image Embedding

This paper presents the 3rd place solution to the Google Universal Image...
research
10/17/2022

2nd Place Solution to Google Universal Image Embedding

Image representations are a critical building block of computer vision a...
research
10/18/2022

5th Place Solution to Kaggle Google Universal Image Embedding Competition

In this paper, we present our solution, which placed 5th in the kaggle G...
research
10/16/2022

1st Place Solution in Google Universal Images Embedding

This paper presents the 1st place solution for the Google Universal Imag...
research
10/08/2021

2nd Place Solution to Google Landmark Retrieval 2021

This paper presents the 2nd place solution to the Google Landmark Retrie...
research
09/02/2019

Enriching Medcial Terminology Knowledge Bases via Pre-trained Language Model and Graph Convolutional Network

Enriching existing medical terminology knowledge bases (KBs) is an impor...
research
06/11/2020

G5: A Universal GRAPH-BERT for Graph-to-Graph Transfer and Apocalypse Learning

The recent GRAPH-BERT model introduces a new approach to learning graph ...

Please sign up or login with your details

Forgot password? Click here to reset