Blockout: Dynamic Model Selection for Hierarchical Deep Networks

12/16/2015
by   Calvin Murdock, et al.
0

Most deep architectures for image classification--even those that are trained to classify a large number of diverse categories--learn shared image representations with a single model. Intuitively, however, categories that are more similar should share more information than those that are very different. While hierarchical deep networks address this problem by learning separate features for subsets of related categories, current implementations require simplified models using fixed architectures specified via heuristic clustering methods. Instead, we propose Blockout, a method for regularization and model selection that simultaneously learns both the model architecture and parameters. A generalization of Dropout, our approach gives a novel parametrization of hierarchical architectures that allows for structure learning via back-propagation. To demonstrate its utility, we evaluate Blockout on the CIFAR and ImageNet datasets, demonstrating improved classification accuracy, better regularization performance, faster training, and the clear emergence of hierarchical network structures.

READ FULL TEXT

page 6

page 12

page 13

page 14

research
06/11/2018

Dual Pattern Learning Networks by Empirical Dual Prediction Risk Minimization

Motivated by the observation that humans can learn patterns from two giv...
research
09/15/2017

Joint Hierarchical Category Structure Learning and Large-Scale Image Classification

We investigate the scalable image classification problem with a large nu...
research
01/05/2021

AutoDropout: Learning Dropout Patterns to Regularize Deep Networks

Neural networks are often over-parameterized and hence benefit from aggr...
research
09/11/2019

Deep Elastic Networks with Model Selection for Multi-Task Learning

In this work, we consider the problem of instance-wise dynamic network m...
research
11/30/2021

Hierarchical clustering: visualization, feature importance and model selection

We propose methods for the analysis of hierarchical clustering that full...
research
05/22/2018

Parsimonious Bayesian deep networks

Combining Bayesian nonparametrics and a forward model selection strategy...
research
11/25/2022

The smooth output assumption, and why deep networks are better than wide ones

When several models have similar training scores, classical model select...

Please sign up or login with your details

Forgot password? Click here to reset