Jointly Image Topic and Emotion Detection using Multi-Modal Hierarchical Latent Dirichlet Allocation

  • Ding, Wanying (College of Computing and Informatics, Drexel University) ;
  • Zhu, Junhuan (Department of Computer Science, University of Rochester) ;
  • Guo, Lifan (TCL Research America) ;
  • Hu, Xiaohua (College of Computing and Informatics, Drexel University) ;
  • Luo, Jiebo (Department of Computer Science, University of Rochester) ;
  • Wang, Haohong (TCL Research America)
  • 투고 : 2014.08.02
  • 심사 : 2014.09.20
  • 발행 : 2014.09.30

초록

Image topic and emotion analysis is an important component of online image retrieval, which nowadays has become very popular in the widely growing social media community. However, due to the gaps between images and texts, there is very limited work in literature to detect one image's Topics and Emotions in a unified framework, although topics and emotions are two levels of semantics that often work together to comprehensively describe one image. In this work, a unified model, Joint Topic/Emotion Multi-Modal Hierarchical Latent Dirichlet Allocation (JTE-MMHLDA) model, which extends previous LDA, mmLDA, and JST model to capture topic and emotion information at the same time from heterogeneous data, is proposed. Specifically, a two level graphical structured model is built to realize sharing topics and emotions among the whole document collection. The experimental results on a Flickr dataset indicate that the proposed model efficiently discovers images' topics and emotions, and significantly outperform the text-only system by 4.4%, vision-only system by 18.1% in topic detection, and outperforms the text-only system by 7.1%, vision-only system by 39.7% in emotion detection.

키워드