Document Domain Randomization for Deep Learning Document Layout Extraction

2Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present document domain randomization (DDR), the first successful transfer of CNNs trained only on graphically rendered pseudo-paper pages to real-world document segmentation. DDR renders pseudo-document pages by modeling randomized textual and non-textual contents of interest, with user-defined layout and font styles to support joint learning of fine-grained classes. We demonstrate competitive results using our DDR approach to extract nine document classes from the benchmark CS-150 and papers published in two domains, namely annual meetings of Association for Computational Linguistics (ACL) and IEEE Visualization (VIS). We compare DDR to conditions of style mismatch, fewer or more noisy samples that are more easily obtained in the real world. We show that high-fidelity semantic information is not necessary to label semantic classes but style mismatch between train and test can lower model accuracy. Using smaller training samples had a slightly detrimental effect. Finally, network models still achieved high test accuracy when correct labels are diluted towards confusing labels; this behavior hold across several classes.

Cite

CITATION STYLE

APA

Ling, M., Chen, J., Möller, T., Isenberg, P., Isenberg, T., Sedlmair, M., … Giles, C. L. (2021). Document Domain Randomization for Deep Learning Document Layout Extraction. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12821 LNCS, pp. 497–513). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-86549-8_32

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free