Deep Mixture of Diverse Experts for Large-Scale Visual Recognition

06/24/2017
by   Tianyi Zhao, et al.
0

In this paper, a deep mixture of diverse experts algorithm is developed for seamlessly combining a set of base deep CNNs (convolutional neural networks) with diverse outputs (task spaces), e.g., such base deep CNNs are trained to recognize different subsets of tens of thousands of atomic object classes. First, a two-layer (category layer and object class layer) ontology is constructed to achieve more effective solution for task group generation, e.g., assigning the semantically-related atomic object classes at the sibling leaf nodes into the same task group because they may share similar learning complexities. Second, one particular base deep CNNs with M+1 (M ≤ 1,000) outputs is learned for each task group to recognize its M atomic object classes effectively and identify one special class of "not-in-group" automatically, and the network structure (numbers of layers and units in each layer) of the well-designed AlexNet is directly used to configure such base deep CNNs. A deep multi-task learning algorithm is developed to leverage the inter-class visual similarities to learn more discriminative base deep CNNs and multi-task softmax for enhancing the separability of the atomic object classes in the same task group. Finally, all these base deep CNNs with diverse outputs (task spaces) are seamlessly combined to form a deep mixture of diverse experts for recognizing tens of thousands of atomic object classes. Our experimental results have demonstrated that our deep mixture of diverse experts algorithm can achieve very competitive results on large-scale visual recognition.

READ FULL TEXT

page 3

page 8

page 9

page 10

research
07/08/2017

Embedding Visual Hierarchy with Deep Networks for Large-Scale Visual Recognition

In this paper, a level-wise mixture model (LMM) is developed by embeddin...
research
10/03/2014

HD-CNN: Hierarchical Deep Convolutional Neural Network for Large Scale Visual Recognition

In image classification, visual separability between different object ca...
research
04/20/2016

Network of Experts for Large-Scale Image Categorization

We present a tree-structured network architecture for large scale image ...
research
08/17/2021

Multi-task learning for jersey number recognition in Ice Hockey

Identifying players in sports videos by recognizing their jersey numbers...
research
08/04/2022

Towards Understanding Mixture of Experts in Deep Learning

The Mixture-of-Experts (MoE) layer, a sparsely-activated model controlle...
research
06/07/2023

Patch-level Routing in Mixture-of-Experts is Provably Sample-efficient for Convolutional Neural Networks

In deep learning, mixture-of-experts (MoE) activates one or few experts ...
research
12/20/2014

Self-informed neural network structure learning

We study the problem of large scale, multi-label visual recognition with...

Please sign up or login with your details

Forgot password? Click here to reset