Virtually Redying Histological Images with Generative Adversarial Networks to Facilitate Unsupervised Segmentation: A Proof-of-Concept Study

0Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Approaches relying on adversarial networks facilitate image-to-image-translation based on unpaired training and thereby open new possibilities for special tasks in image analysis. We propose a methodology to improve segmentability of histological images by making use of image-to-image translation. We generate virtual stains and exploit the additional information during segmentation. Specifically a very basic pixel-based segmentation approach is applied in order to focus on the information content available on pixel-level and to avoid any bias which might be introduced by more elaborated techniques. The results of this proof-of-concept trial indicate a performance gain compared to segmentation with the source stain only. Further experiments including more powerful supervised state-of-the-art machine learning approaches and larger evaluation data sets need to follow.

Cite

CITATION STYLE

APA

Gadermayr, M., Klinkhammer, B. M., & Boor, P. (2019). Virtually Redying Histological Images with Generative Adversarial Networks to Facilitate Unsupervised Segmentation: A Proof-of-Concept Study. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11435 LNCS, pp. 38–46). Springer Verlag. https://doi.org/10.1007/978-3-030-23937-4_5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free