TabTransformer: Tabular Data Modeling Using Contextual Embeddings

12/11/2020
by   Xin Huang, et al.
8

We propose TabTransformer, a novel deep tabular data modeling architecture for supervised and semi-supervised learning. The TabTransformer is built upon self-attention based Transformers. The Transformer layers transform the embeddings of categorical features into robust contextual embeddings to achieve higher prediction accuracy. Through extensive experiments on fifteen publicly available datasets, we show that the TabTransformer outperforms the state-of-the-art deep learning methods for tabular data by at least 1.0 mean AUC, and matches the performance of tree-based ensemble models. Furthermore, we demonstrate that the contextual embeddings learned from TabTransformer are highly robust against both missing and noisy data features, and provide better interpretability. Lastly, for the semi-supervised setting we develop an unsupervised pre-training procedure to learn data-driven contextual embeddings, resulting in an average 2.1 methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/09/2021

Lexico-semantic and affective modelling of Spanish poetry: A semi-supervised learning approach

Text classification tasks have improved substantially during the last ye...
research
04/08/2021

Embeddings and Attention in Predictive Modeling

We explore in depth how categorical data can be processed with embedding...
research
01/04/2023

Semi-MAE: Masked Autoencoders for Semi-supervised Vision Transformers

Vision Transformer (ViT) suffers from data scarcity in semi-supervised l...
research
06/05/2020

DeCLUTR: Deep Contrastive Learning for Unsupervised Textual Representations

We present DeCLUTR: Deep Contrastive Learning for Unsupervised Textual R...
research
12/03/2021

CTIN: Robust Contextual Transformer Network for Inertial Navigation

Recently, data-driven inertial navigation approaches have demonstrated t...
research
05/11/2020

CrisisBERT: Robust Transformer for Crisis Classification and Contextual Crisis Embedding

Classification of crisis events, such as natural disasters, terrorist at...
research
05/11/2020

CrisisBERT: a Robust Transformer for Crisis Classification and Contextual Crisis Embedding

Classification of crisis events, such as natural disasters, terrorist at...

Please sign up or login with your details

Forgot password? Click here to reset