Learning to segment medical images with scribble-supervision alone

N/ACitations
Citations of this article
75Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Semantic segmentation of medical images is a crucial step for the quantification of healthy anatomy and diseases alike. The majority of the current state-of-the-art segmentation algorithms are based on deep neural networks and rely on large datasets with full pixel-wise annotations. Producing such annotations can often only be done by medical professionals and requires large amounts of valuable time. Training a medical image segmentation network with weak annotations remains a relatively unexplored topic. In this work we investigate training strategies to learn the parameters of a pixel-wise segmentation network from scribble annotations alone. We evaluate the techniques on public cardiac (ACDC) and prostate (NCI-ISBI) segmentation datasets. We find that the networks trained on scribbles suffer from a remarkably small degradation in Dice of only 2.9% (cardiac) and 4.5% (prostate) with respect to a network trained on full annotations.

Cite

CITATION STYLE

APA

Can, Y. B., Chaitanya, K., Mustafa, B., Koch, L. M., Konukoglu, E., & Baumgartner, C. F. (2018). Learning to segment medical images with scribble-supervision alone. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11045 LNCS, pp. 236–244). Springer Verlag. https://doi.org/10.1007/978-3-030-00889-5_27

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free