Wang, Cheng; Yang, Haojin; Che, Xiaoyin; Meinel, Christoph
Proceedings of the 21st International Conference on Multimedia Modeling (MMM2015)
Lecture Notes in Computer Science
In this paper, we propose a concept-based multimodal learning model (CMLM) for generating document topic through modeling textual and visual data. Our model considers cross-modal concept similarity and unlabeled image concept, it is capable of processing document which has modality missing. The model can extract semantic concepts from unlabeled image and combine with text modality to generate document topics. Our comparison experiments on news document topic generation shows, in multimodal scenario, CMLM can generate more representative topics than latent dirichet allocation (LDA) based topic for representing given document.