STRIVE: Scene Text Replacement In Videos

09/06/2021
by   Vijay Kumar B G, et al.
0

We propose replacing scene text in videos using deep style transfer and learned photometric transformations.Building on recent progress on still image text replacement,we present extensions that alter text while preserving the appearance and motion characteristics of the original video.Compared to the problem of still image text replacement,our method addresses additional challenges introduced by video, namely effects induced by changing lighting, motion blur, diverse variations in camera-object pose over time,and preservation of temporal consistency. We parse the problem into three steps. First, the text in all frames is normalized to a frontal pose using a spatio-temporal trans-former network. Second, the text is replaced in a single reference frame using a state-of-art still-image text replacement method. Finally, the new text is transferred from the reference to remaining frames using a novel learned image transformation network that captures lighting and blur effects in a temporally consistent manner. Results on synthetic and challenging real videos show realistic text trans-fer, competitive quantitative and qualitative performance,and superior inference speed relative to alternatives. We introduce new synthetic and real-world datasets with paired text objects. To the best of our knowledge this is the first attempt at deep video text replacement.

READ FULL TEXT

page 1

page 6

page 7

page 8

research
11/25/2016

Deep Video Deblurring

Motion blur from camera shake is a major problem in videos captured by h...
research
08/26/2019

Mocycle-GAN: Unpaired Video-to-Video Translation

Unsupervised image-to-image translation is the task of translating an im...
research
06/07/2020

Realistic text replacement with non-uniform style conditioning

In this work, we study the possibility of realistic text replacement, th...
research
06/24/2019

Deep Exemplar-based Video Colorization

This paper presents the first end-to-end network for exemplar-based vide...
research
05/30/2023

Video ControlNet: Towards Temporally Consistent Synthetic-to-Real Video Translation Using Conditional Image Diffusion Models

In this study, we present an efficient and effective approach for achiev...
research
07/06/2023

Synthesizing Artistic Cinemagraphs from Text

We introduce Text2Cinemagraph, a fully automated method for creating cin...
research
09/06/2023

Synthetic Text Generation using Hypergraph Representations

Generating synthetic variants of a document is often posed as text-to-te...

Please sign up or login with your details

Forgot password? Click here to reset