Mediated Experts for Deep Convolutional Networks

11/19/2015
by   Sebastian Agethen, et al.
0

We present a new supervised architecture termed Mediated Mixture-of-Experts (MMoE) that allows us to improve classification accuracy of Deep Convolutional Networks (DCN). Our architecture achieves this with the help of expert networks: A network is trained on a disjoint subset of a given dataset and then run in parallel to other experts during deployment. A mediator is employed if experts contradict each other. This allows our framework to naturally support incremental learning, as adding new classes requires (re-)training of the new expert only. We also propose two measures to control computational complexity: An early-stopping mechanism halts experts that have low confidence in their prediction. The system allows to trade-off accuracy and complexity without further retraining. We also suggest to share low-level convolutional layers between experts in an effort to avoid computation of a near-duplicate feature set. We evaluate our system on a popular dataset and report improved accuracy compared to a single model of same configuration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/05/2018

Deep Mixture of Experts via Shallow Embedding

Larger networks generally have greater representational power at the cos...
research
04/20/2016

Network of Experts for Large-Scale Image Categorization

We present a tree-structured network architecture for large scale image ...
research
08/09/2023

Expert load matters: operating networks at high accuracy and low manual effort

In human-AI collaboration systems for critical applications, in order to...
research
06/06/2016

Very Deep Convolutional Networks for Text Classification

The dominant approach for many NLP tasks are recurrent neural networks, ...
research
12/11/2019

Robust Gabor Networks

This work takes a step towards investigating the benefits of merging cla...
research
04/20/2017

Hard Mixtures of Experts for Large Scale Weakly Supervised Vision

Training convolutional networks (CNN's) that fit on a single GPU with mi...
research
11/26/2017

SkipNet: Learning Dynamic Routing in Convolutional Networks

Increasing depth and complexity in convolutional neural networks has ena...

Please sign up or login with your details

Forgot password? Click here to reset