Yunzhi LingYing WangXin WangYunhao Ling
In multi-label learning, instances can be associated with a set of class labels. The existing multi-label feature selection (MLFS) methods generally adopt either of these two strategies, namely, selecting a subset of features that is shared by all labels (common features) or exploring the most discriminative features for each label (label-specific features). However, both of them can play a key role in the discrimination of different labels. For example, common features can distinguish all labels, and label-specific features contribute to discriminating label’s differences. They are important for the discriminability of selected features. On the other hand, it is well-known that exploiting label correlations can advance the performance of MLFS, and label correlations are local and only shared by a data subset in most cases. How to effectively learn and exploit local label correlations in the selection process is significant. In this paper, to address these problems, we propose a novel MLFS framework. Specially, common and label-specific features are simultaneously considered by introducing both $l_{2,1}$ -norm and $l_{1}$ -norm regularizers, local label correlations are automatically learned with probability and learned correlation information is efficiently exploited to help feature selection by constraining label correlations on the output of labels. A comparative study with seven state-of-the-art methods manifests the efficacy of our framework.
Dawei ZhaoQingwei GaoYixiang LuDong Sun
Jia ZhangCandong LiDonglin CaoYaojin LinSongzhi SuLiang DaiShaozi Li
Wei WengBowen WeiKe WenYuling FanJinbo WangYuwen Li
Mengxuan SunPeipei LiJunlong LiXuegang Hu