ExcelFormer: A Neural Network Surpassing GBDTs on Tabular Data

01/07/2023
by   Jintai Chen, et al.
0

Though neural networks have achieved enormous breakthroughs on various fields (e.g., computer vision) in supervised learning, they still trailed the performances of GBDTs on tabular data thus far. Delving into this issue, we identify that a proper handling of feature interactions and feature embedding is crucial to the success of neural networks on tabular data. We develop a novel neural network called ExcelFormer, which alternates in turn two attention modules that respectively manipulate careful feature interactions and feature embedding updates. A bespoke training methodology is jointly introduced to facilitate the model performances. By initializing parameters with minuscule values, these attention modules are attenuated when the training begins, and the effects of feature interactions and embedding updates progressively grow up to optimum levels under the guidance of the proposed specific regularization approaches Swap-Mix and Hidden-Mix as the training proceeds. Experiments on 25 public tabular datasets show that our ExcelFormer is superior to extremely-tuned GBDTs, which is an unprecedented achievement of neural networks in supervised tabular learning.

READ FULL TEXT

page 4

page 8

research
08/01/2022

Locally Supervised Learning with Periodic Global Guidance

Locally supervised learning aims to train a neural network based on a lo...
research
07/06/2020

GateNet: Gating-Enhanced Deep Network for Click-Through Rate Prediction

Advertising and feed ranking are essential to many Internet companies su...
research
10/08/2016

Revisiting Multiple Instance Neural Networks

Recently neural networks and multiple instance learning are both attract...
research
06/22/2022

Consistency of Neural Networks with Regularization

Neural networks have attracted a lot of attention due to its success in ...
research
12/26/2018

Learning Not to Learn: Training Deep Neural Networks with Biased Data

We propose a novel regularization algorithm to train deep neural network...
research
07/13/2020

Deep Reinforced Attention Learning for Quality-Aware Visual Recognition

In this paper, we build upon the weakly-supervised generation mechanism ...

Please sign up or login with your details

Forgot password? Click here to reset