The Contextual Loss for Image Transformation with Non-Aligned Data

03/06/2018 ∙ by Roey Mechrez, et al. ∙ 0

Feed-forward CNNs trained for image transformation problems rely on loss functions that measure the similarity between the generated image and a target image. Most of the common loss functions assume that these images are spatially aligned and compare pixels at corresponding locations. However, for many tasks, aligned training pairs of images will not be available. We present an alternative loss function that does not require alignment, thus providing an effective and simple solution for a new space of problems. Our loss is based on both context and semantics -- it compares regions with similar semantic meaning, while considering the context of the entire image. Hence, for example, when transferring the style of one face to another, it will translate eyes-to-eyes and mouth-to-mouth.

READ FULL TEXT

Authors

page 1

page 2

page 8

page 10

page 11

page 12

page 13

page 14

Code Repositories

contextualLoss

The Contextual Loss


view repo
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.