Categorical Representation Learning: Morphism is All You Need

03/26/2021
by   Artan Sheshmani, et al.
0

We provide a construction for categorical representation learning and introduce the foundations of "categorifier". The central theme in representation learning is the idea of everything to vector. Every object in a dataset 𝒮 can be represented as a vector in ℝ^n by an encoding map E: 𝒪bj(𝒮)→ℝ^n. More importantly, every morphism can be represented as a matrix E: ℋom(𝒮)→ℝ^n_n. The encoding map E is generally modeled by a deep neural network. The goal of representation learning is to design appropriate tasks on the dataset to train the encoding map (assuming that an encoding is optimal if it universally optimizes the performance on various tasks). However, the latter is still a set-theoretic approach. The goal of the current article is to promote the representation learning to a new level via a category-theoretic approach. As a proof of concept, we provide an example of a text translator equipped with our technology, showing that our categorical learning model outperforms the current deep learning models by 17 times. The content of the current article is part of the recent US patent proposal (patent application number: 63110906).

READ FULL TEXT

page 2

page 13

research
05/25/2022

NECA: Network-Embedded Deep Representation Learning for Categorical Data

We propose NECA, a deep representation learning method for categorical d...
research
12/22/2021

MC-DGCNN: A Novel DNN Architecture for Multi-Category Point Set Classification

Point set classification aims to build a representation learning model t...
research
03/15/2022

Categorical Representation Learning and RG flow operators for algorithmic classifiers

Following the earlier formalism of the categorical representation learni...
research
02/15/2022

On Representation Learning with Feedback

This note complements the author's recent paper "Robust representation l...
research
07/21/2020

Unsupervised Heterogeneous Coupling Learning for Categorical Representation

Complex categorical data is often hierarchically coupled with heterogene...
research
11/05/2022

Small Language Models for Tabular Data

Supervised deep learning is most commonly applied to difficult problems ...
research
05/07/2021

GADTs, Functoriality, Parametricity: Pick Two

GADTs can be represented either as their Church encodings à la Atkey, or...

Please sign up or login with your details

Forgot password? Click here to reset