Self-supervised Document Clustering Based on BERT with Data Augment

11/17/2020
by   Haoxiang Shi, et al.
0

Contrastive learning is a good way to pursue discriminative unsupervised learning, which can inherit advantages and experiences of well-studied deep models without complexly novel model designing. In this paper, we propose two learning method for document clustering, the one is a partial contrastive learning with unsupervised data augment, and the other is a self-supervised contrastive learning. Both methods achieve state-of-the-art results in clustering accuracy when compared to recently proposed unsupervised clustering approaches.

READ FULL TEXT
research
03/15/2022

SCD: Self-Contrastive Decorrelation for Sentence Embeddings

In this paper, we propose Self-Contrastive Decorrelation (SCD), a self-s...
research
04/19/2023

Shuffle Divide: Contrastive Learning for Long Text

We propose a self-supervised learning method for long text documents bas...
research
03/30/2021

Contrastive Learning of Single-Cell Phenotypic Representations for Treatment Classification

Learning robust representations to discriminate cell phenotypes based on...
research
07/27/2020

Self-Supervised Contrastive Learning for Unsupervised Phoneme Segmentation

We propose a self-supervised representation learning model for the task ...
research
08/05/2021

A Low Rank Promoting Prior for Unsupervised Contrastive Learning

Unsupervised learning is just at a tipping point where it could really t...
research
08/01/2019

Contrastive Reasons Detection and Clustering from Online Polarized Debate

This work tackles the problem of unsupervised modeling and extraction of...
research
02/19/2023

Supervised Contrastive Learning and Feature Fusion for Improved Kinship Verification

Facial Kinship Verification is the task of determining the degree of fam...

Please sign up or login with your details

Forgot password? Click here to reset