会议专题

Latent Gaussian-Multinomial Generative Model for Annotated Data

  Traditional generative models annotate images by multiple instances independently segmented,but these models have been becoming prohibitively expensive and time-consuming along with the growth of Internet data.Focusing on the annotated data,we propose a latent Gaussian-Multinomial generative model(LGMG),which generates the image-annotations using a multimodal probabilistic models.Specifically,we use a continuous latent variable with prior of Normal distribution as the latent representation summarizing the high-level semantics of images,and a discrete latent variable with prior of Multinomial distribution as the topics indicator for annotation.We compute the variational posteriors from a mapping structure among latent representation,topics indicator and image-annotation.The stochastic gradient variational Bayes estimator on variational objective is realized by combining the reparameterization trick and Monte Carlo estimator.Finally,we demonstrate the performance of LGMG on LabelMe in terms of held-out likelihood,automatic image annotation with the state-of-the-art models.

Annotated data Gaussian-Multinomial Multimodal generative models Latent representation Topics indicator

Shuoran Jiang Yarui Chen Zhifei Qin Jucheng Yang Tingting Zhao Chuanlei Zhang

Tianjin University of Science and Technology,Tianjin 300457,China

国际会议

The 23rd Pacific-Asia Conference on Knowledge Discovery and Data Mining (第23届亚太知识发现和数据挖掘国际会议(PAKDD2019)

澳门

英文

42-54

2019-04-14(万方平台首次上网日期,不代表论文的发表时间)