MS-DINO: Efficient Distributed Training of Vision Transformer Foundation Model in Medical Domain through Masked Sampling

01/05/2023
by   Sangjoon Park, et al.
0

In spite of the recent success of deep learning in the medical domain, the problem of data scarcity in the medical domain gets aggravated due to privacy and data ownership issues. Distributed learning approaches including federated learning have been studied to alleviate the problems, but they suffer from cumbersome communication overheads and weakness in privacy protection. To address this, here we propose a self-supervised masked sampling distillation method for vision transformer that can be performed without continuous communication but still enhance privacy using a vision transformer-specific encryption method. The effectiveness of our method is demonstrated with extensive experiments on two medical domain data and two different downstream tasks, showing superior performances than those obtained with the existing distributed learning strategy as well as the fine-tuning only baseline. As the self-supervised model built with the proposed method is capable of having a general semantic understanding of the modality, we demonstrate its potential as a task-agnostic foundation model for various medical tasks, widening the applicability in the medical domain.

READ FULL TEXT

page 3

page 4

page 8

page 10

page 11

research
07/04/2023

SelfFed: Self-supervised Federated Learning for Data Heterogeneity and Label Scarcity in IoMT

Self-supervised learning in federated learning paradigm has been gaining...
research
09/25/2022

Dive into Self-Supervised Learning for Medical Image Analysis: Data, Models and Tasks

Self-supervised learning (SSL) has achieved remarkable performance on va...
research
04/05/2023

Towards Efficient Task-Driven Model Reprogramming with Foundation Models

Vision foundation models exhibit impressive power, benefiting from the e...
research
09/29/2021

Federated Self-Supervised Contrastive Learning via Ensemble Similarity Distillation

This paper investigates the feasibility of learning good representation ...
research
05/19/2023

Federated Foundation Models: Privacy-Preserving and Collaborative Learning for Large Models

Foundation Models (FMs), such as BERT, GPT, ViT, and CLIP, have demonstr...
research
04/07/2022

Multi-Task Distributed Learning using Vision Transformer with Random Patch Permutation

The widespread application of artificial intelligence in health research...
research
03/08/2022

SuperPoint features in endoscopy

There is often a significant gap between research results and applicabil...

Please sign up or login with your details

Forgot password? Click here to reset