Partial multi-label learning using label compression

Tingting Yu, Guoxian Yu, Jun Wang, Carlotta Domeniconi, Xiangliang Zhang

Research output: Chapter in Book/Report/Conference proceedingConference contribution

14 Scopus citations


Partial multi-label learning (PML) aims at learning a robust multi-label classifier from partial multi-label data, where a sample is annotated with a set of candidate labels, while only a subset of those labels is valid. The existing PML algorithms generally suffer from the high computational cost when learning with large label spaces. In this paper, we introduce a PML approach (PML-LCom) that uses Label Compression to efficiently learn from partial multi-label data. PML-LCom firstly splits the observed label data matrix into a latent relevant label matrix and an irrelevant one, and then factorizes the relevant label matrix into two low-rank matrices, one encodes the compressed labels of samples, and the other explores the underlying label correlations. Next, it optimizes the coefficient matrix of the multi-label predictor with respect to the compressed label matrix. In addition, it regularizes the compressed label matrix with respect to the feature similarity of samples, and optimizes the label matrix and predictor in a coherent manner. Experimental results on both semi-synthetic and real-world PML datasets show that PML-LCom achieves a performance superior to the state-of-the-art solutions on predicting the labels of unlabeled samples with a large label space. The label compression improves both the effectiveness and efficiency, and the coherent optimization mutually benefits the label matrix and predictor.
Original languageEnglish (US)
Title of host publication2020 IEEE International Conference on Data Mining (ICDM)
Number of pages10
ISBN (Print)9781728183169
StatePublished - Nov 2020


Dive into the research topics of 'Partial multi-label learning using label compression'. Together they form a unique fingerprint.

Cite this