Log In Sign Up

Epoch-wise label attacks for robustness against label noise

by   Sebastian Guendel, et al.

The current accessibility to large medical datasets for training convolutional neural networks is tremendously high. The associated dataset labels are always considered to be the real "groundtruth". However, the labeling procedures often seem to be inaccurate and many wrong labels are integrated. This may have fatal consequences on the performance on both training and evaluation. In this paper, we show the impact of label noise in the training set on a specific medical problem based on chest X-ray images. With a simple one-class problem, the classification of tuberculosis, we measure the performance on a clean evaluation set when training with label-corrupt data. We develop a method to compete with incorrectly labeled data during training by randomly attacking labels on individual epochs. The network tends to be robust when flipping correct labels for a single epoch and initiates a good step to the optimal minimum on the error surface when flipping noisy labels. On a baseline with an AUC (Area under Curve) score of 0.924, the performance drops to 0.809 when 30 our approach the baseline performance could almost be maintained, the performance raised to 0.918.


Noisy Label Learning for Large-scale Medical Image Classification

The classification accuracy of deep learning models depends not only on ...

Leveraging inductive bias of neural networks for learning without explicit human annotations

Classification problems today are typically solved by first collecting e...

ContamiNet: Detecting Contamination in Municipal Solid Waste

Leveraging over 30,000 images each with up to 89 labels collected by Rec...

Exploiting Context for Robustness to Label Noise in Active Learning

Several works in computer vision have demonstrated the effectiveness of ...

Making Every Label Count: Handling Semantic Imprecision by Integrating Domain Knowledge

Noisy data, crawled from the web or supplied by volunteers such as Mecha...

Explainability Matters: Backdoor Attacks on Medical Imaging

Deep neural networks have been shown to be vulnerable to backdoor attack...