CAST: Character labeling in Animation using Self-supervision by Tracking

01/19/2022
by   Oron Nir, et al.
3

Cartoons and animation domain videos have very different characteristics compared to real-life images and videos. In addition, this domain carries a large variability in styles. Current computer vision and deep-learning solutions often fail on animated content because they were trained on natural images. In this paper we present a method to refine a semantic representation suitable for specific animated content. We first train a neural network on a large-scale set of animation videos and use the mapping to deep features as an embedding space. Next, we use self-supervision to refine the representation for any specific animation style by gathering many examples of animated characters in this style, using a multi-object tracking. These examples are used to define triplets for contrastive loss training. The refined semantic space allows better clustering of animated characters even when they have diverse manifestations. Using this space we can build dictionaries of characters in an animation videos, and define specialized classifiers for specific stylistic content (e.g., characters in a specific animation series) with very little user effort. These classifiers are the basis for automatically labeling characters in animation videos. We present results on a collection of characters in a variety of animation styles.

READ FULL TEXT

page 2

page 3

page 6

page 15

page 16

page 17

page 18

page 19

research
08/25/2020

Multi-Face: Self-supervised Multiview Adaptation for Robust Face Clustering in Videos

Robust face clustering is a key step towards computational understanding...
research
05/26/2020

CalliGAN: Style and Structure-aware Chinese Calligraphy Character Generator

Chinese calligraphy is the writing of Chinese characters as an art form ...
research
07/27/2023

PromptStyler: Prompt-driven Style Generation for Source-free Domain Generalization

In a joint vision-language space, a text feature (e.g., from "a photo of...
research
10/28/2019

FontGAN: A Unified Generative Framework for Chinese Character Stylization and De-stylization

Chinese character synthesis involves two related aspects, i.e., style ma...
research
03/27/2023

Handwritten Text Generation from Visual Archetypes

Generating synthetic images of handwritten text in a writer-specific sty...
research
03/26/2023

Disentangling Writer and Character Styles for Handwriting Generation

Training machines to synthesize diverse handwritings is an intriguing ta...
research
04/15/2020

Visual Descriptor Learning from Monocular Video

Correspondence estimation is one of the most widely researched and yet o...

Please sign up or login with your details

Forgot password? Click here to reset