In this paper, we propose a concept-based multimodal learning model (CMLM) for generating document topic through modeling textual and visual data. Our model considers cross-modal concept similarity and unlabeled image concept, it is capable of processing document which has modality missing. The model can extract semantic concepts from unlabeled image and combine with text modality to generate document topics. Our comparison experiments on news document topic generation shows, in multimodal scenario, CMLM can generate more representative topics than latent dirichet allocation (LDA) based topic for representing given document.
CITATION STYLE
Wang, C., Yang, H., Che, X., & Meinel, C. (2015). Concept-based multimodal learning for topic generation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8935, pp. 385–395). Springer Verlag. https://doi.org/10.1007/978-3-319-14445-0_33
Mendeley helps you to discover research relevant for your work.