Exploring Common and Label-Specific Features for Multi-Label Learning With Local Label Correlations

In multi-label learning, instances can be associated with a set of class labels. The existing multi-label feature selection (MLFS) methods generally adopt either of these two strategies, namely, selecting a subset of features that is shared by all labels (common features) or exploring the most discr...

Full description

Bibliographic Details
Main Authors: Yunzhi Ling, Ying Wang, Xin Wang, Yunhao Ling
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9032184/
Description
Summary:In multi-label learning, instances can be associated with a set of class labels. The existing multi-label feature selection (MLFS) methods generally adopt either of these two strategies, namely, selecting a subset of features that is shared by all labels (common features) or exploring the most discriminative features for each label (label-specific features). However, both of them can play a key role in the discrimination of different labels. For example, common features can distinguish all labels, and label-specific features contribute to discriminating label's differences. They are important for the discriminability of selected features. On the other hand, it is well-known that exploiting label correlations can advance the performance of MLFS, and label correlations are local and only shared by a data subset in most cases. How to effectively learn and exploit local label correlations in the selection process is significant. In this paper, to address these problems, we propose a novel MLFS framework. Specially, common and label-specific features are simultaneously considered by introducing both $l_{2,1}$ -norm and $l_{1}$ -norm regularizers, local label correlations are automatically learned with probability and learned correlation information is efficiently exploited to help feature selection by constraining label correlations on the output of labels. A comparative study with seven state-of-the-art methods manifests the efficacy of our framework.
ISSN:2169-3536