On Random Weights for Texture Generation in One Layer Neural Networks

12/19/2016
by   Mihir Mongia, et al.
0

Recent work in the literature has shown experimentally that one can use the lower layers of a trained convolutional neural network (CNN) to model natural textures. More interestingly, it has also been experimentally shown that only one layer with random filters can also model textures although with less variability. In this paper we ask the question as to why one layer CNNs with random filters are so effective in generating textures? We theoretically show that one layer convolutional architectures (without a non-linearity) paired with the an energy function used in previous literature, can in fact preserve and modulate frequency coefficients in a manner so that random weights and pretrained weights will generate the same type of images. Based on the results of this analysis we question whether similar properties hold in the case where one uses one convolution layer with a non-linearity. We show that in the case of ReLu non-linearity there are situations where only one input will give the minimum possible energy whereas in the case of no nonlinearity, there are always infinite solutions that will give the minimum possible energy. Thus we can show that in certain situations adding a ReLu non-linearity generates less variable images.

READ FULL TEXT
research
05/31/2016

Texture Synthesis Using Shallow Convolutional Networks with Random Filters

Here we demonstrate that the feature space of random shallow convolution...
research
02/22/2017

Synthesising Dynamic Textures using Convolutional Neural Networks

Here we present a parametric model for dynamic textures. The model is ba...
research
06/07/2022

Adversarial Reprogramming Revisited

Adversarial reprogramming, introduced by Elsayed, Goodfellow, and Sohl-D...
research
01/08/2019

Interpretable CNNs

This paper proposes a generic method to learn interpretable convolutiona...
research
06/15/2018

Random depthwise signed convolutional neural networks

Random weights in convolutional neural networks have shown promising res...
research
05/27/2021

Passing Multi-Channel Material Textures to a 3-Channel Loss

Our objective is to compute a textural loss that can be used to train te...

Please sign up or login with your details

Forgot password? Click here to reset