Deep Distributed Random Samplings for Supervised Learning: An Alternative to Random Forests?

12/03/2014
by   Xiao-Lei Zhang, et al.
0

In (zhang2014nonlinear,zhang2014nonlinear2), we have viewed machine learning as a coding and dimensionality reduction problem, and further proposed a simple unsupervised dimensionality reduction method, entitled deep distributed random samplings (DDRS). In this paper, we further extend it to supervised learning incrementally. The key idea here is to incorporate label information into the coding process by reformulating that each center in DDRS has multiple output units indicating which class the center belongs to. The supervised learning method seems somewhat similar with random forests (breiman2001random), here we emphasize their differences as follows. (i) Each layer of our method considers the relationship between part of the data points in training data with all training data points, while random forests focus on building each decision tree on only part of training data points independently. (ii) Our method builds gradually-narrowed network by sampling less and less data points, while random forests builds gradually-narrowed network by merging subclasses. (iii) Our method is trained more straightforward from bottom layer to top layer, while random forests build each tree from top layer to bottom layer by splitting. (iv) Our method encodes output targets implicitly in sparse codes, while random forests encode output targets by remembering the class attributes of the activated nodes. Therefore, our method is a simpler, more straightforward, and maybe a better alternative choice, though both methods use two very basic elements---randomization and nearest neighbor optimization---as the core. This preprint is used to protect the incremental idea from (zhang2014nonlinear,zhang2014nonlinear2). Full empirical evaluation will be announced carefully later.

READ FULL TEXT
research
07/05/2022

An Approximation Method for Fitted Random Forests

Random Forests (RF) is a popular machine learning method for classificat...
research
08/05/2014

Multilayer bootstrap networks

Multilayer bootstrap network builds a gradually narrowed multilayer nonl...
research
10/30/2022

Gravitational Dimensionality Reduction Using Newtonian Gravity and Einstein's General Relativity

Due to the effectiveness of using machine learning in physics, it has be...
research
05/15/2019

Cluster, Classify, Regress: A General Method For Learning Discountinous Functions

This paper presents a method for solving the supervised learning problem...
research
02/07/2012

Information Forests

We describe Information Forests, an approach to classification that gene...
research
08/22/2013

Learning Deep Representation Without Parameter Inference for Nonlinear Dimensionality Reduction

Unsupervised deep learning is one of the most powerful representation le...
research
10/29/2020

Analyzing the tree-layer structure of Deep Forests

Random forests on the one hand, and neural networks on the other hand, h...

Please sign up or login with your details

Forgot password? Click here to reset