A closer look at domain shift for deep learning in histopathology

09/25/2019
by   Karin Stacke, et al.
0

Domain shift is a significant problem in histopathology. There can be large differences in data characteristics of whole-slide images between medical centers and scanners, making generalization of deep learning to unseen data difficult. To gain a better understanding of the problem, we present a study on convolutional neural networks trained for tumor classification of H&E stained whole-slide images. We analyze how augmentation and normalization strategies affect performance and learned representations, and what features a trained model respond to. Most centrally, we present a novel measure for evaluating the distance between domains in the context of the learned representation of a particular model. This measure can reveal how sensitive a model is to domain variations, and can be used to detect new data that a model will have problems generalizing to. The results show how learning is heavily influenced by the preparation of training data, and that the latent representation used to do classification is sensitive to changes in data distribution, especially when training without augmentation or normalization.

READ FULL TEXT
research
07/18/2023

Adversarial Bayesian Augmentation for Single-Source Domain Generalization

Generalizing to unseen image domains is a challenging problem primarily ...
research
05/09/2022

Towards Measuring Domain Shift in Histopathological Stain Translation in an Unsupervised Manner

Domain shift in digital histopathology can occur when different stains o...
research
02/18/2019

Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology

Stain variation is a phenomenon observed when distinct pathology laborat...
research
10/15/2020

An Algorithm for Learning Smaller Representations of Models With Scarce Data

We present a greedy algorithm for solving binary classification problems...
research
05/20/2022

Unintended memorisation of unique features in neural networks

Neural networks pose a privacy risk due to their propensity to memorise ...
research
03/03/2023

Unproportional mosaicing

Data shift is a gap between data distribution used for training and data...
research
02/10/2020

Be Like Water: Robustness to Extraneous Variables Via Adaptive Feature Normalization

Extraneous variables are variables that are irrelevant for a certain tas...

Please sign up or login with your details

Forgot password? Click here to reset