Large-scale Bayesian multi-label learning via topic-based label embeddings

Published

Conference Paper

We present a scalable Bayesian multi-label learning model based on learning lowdimensional label embeddings. Our model assumes that each label vector is generated as a weighted combination of a set of topics (each topic being a distribution over labels), where the combination weights (i.e., the embeddings) for each label vector are conditioned on the observed feature vector. This construction, coupled with a Bernoulli-Poisson link function for each label of the binary label vector, leads to a model with a computational cost that scales in the number of positive labels in the label matrix. This makes the model particularly appealing for real-world multi-label learning problems where the label matrix is usually very massive but highly sparse. Using a data-augmentation strategy leads to full local conjugacy in our model, facilitating simple and very efficient Gibbs sampling, as well as an Expectation Maximization algorithm for inference. Also, predicting the label vector at test time does not require doing an inference for the label embeddings and can be done in closed form. We report results on several benchmark data sets, comparing our model with various state-of-the art methods.

Duke Authors

Cited Authors

  • Raiy, P; Hu, C; Henao, R; Carin, L

Published Date

  • January 1, 2015

Published In

Volume / Issue

  • 2015-January /

Start / End Page

  • 3222 - 3230

International Standard Serial Number (ISSN)

  • 1049-5258

Citation Source

  • Scopus