Labeling Information Enhancement for Multi-label Learning with Low-rank Subspace An Tao* , Ning Xu, and Xin Geng Southeast University, China
Outline Introduction 1 The LIEML Algorithm 2 Experiments 3 Conclusion 4
Introduction 1
Introduction In traditional multi-label learning: Features: each pixels in the picture Label set: Y = { sky, water, cloud, beach, plant, house } Labeling information in multi-labellearning is categorical in essence. • Each label is regarded to be either relevant or irrelevant to the instance. We call such label as logical label .
Introduction However… Same logical label set Y = { sky, water, cloud, beach, plant, house } Q: The two pictures can’t be differed with only logical labels. A: To describe the pictures better, we extend the logical label to be numerical. This new label is called numerical label .
Introduction Specification of logical label Use 𝒛 " ∈ {−1,1} ) to denote the logical label vector. Label element of 𝒛 " = 1 : relevant to the instance. • Label element of 𝒛 " = −1 : irrelevant to the instance. • Specification of numerical label Use to denote the numerical label vector. > 0 : relevant to the instance. • Label element of < 0 : irrelevant to the instance. • Label element of • Absolute value of label element of : reflects the degree to which the label describes the instance.
Introduction Overview of our LIEML algorithm for multi-label learning Logical Label Step 1: Label Enhancement (LE) + Numerical Label Feature Step 2: Predictive Model Induction Multi-label Model ü LE can be seen as a data preprocessing step which aims to facilitate in learning a better model.
The LIEML Algorithm 2
The LIEMLAlgorithm Symbol Definition: • : Dataset • : i -th instance vector • : i -th logical label vector • : i -th numerical label vector Label Enhancement Linear Model: • is a weight matrix. • is a bias vector.
The LIEMLAlgorithm Label Enhancement Linear Model: • For convenient describing, we set . The model becomes: We then construct a stacked matrix Z : Label Enhancement ’ Target Matrix
The LIEMLAlgorithm Label Enhancement Not deviate too much ’ The optimization problem of LE becomes: • L ( Z ): logistic loss function ü It prevents the label values in Z from deviating the original values too much.
The LIEMLAlgorithm Labeling Information ↑ Label Enhancement Not deviate too much ’ Rank( Z ) ↓ • L ( Z ): nuclear norm and squared function Low-rank Assumption Q: Why construct the stacked matrix Z? A: We assume that the stacked matrix Z belongs to an underlying low-rank subspace. ü The stacked matrix Z is therefore an underlying low-rank matrix.
The LIEMLAlgorithm Label Enhancement Not deviate too much ’ Rank( Z ) ↓ The target function T 1 for optimization is yielded as:
The LIEMLAlgorithm Predictive Model Induction We build the learning model through an adapted regressor based on MSVR. The target function T 2 we wish to minimize is: • , , , , , , , , . . • is a nonlinear transformation of x to a higher dimensional feature space.
The LIEMLAlgorithm
Experiments 3
Experiments Experiment Configuration Ten benchmark multi-label data sets: Six well-established multi-label learning algorithms: • BR, CLR, ECC, RAKEL, LP, and ML 2 Five evaluation metrics widely-used in multi-label learning: • Ranking-loss, One-error, Hamming-loss, Coverage, and Average precision
Experiments Experimental Results
Experiments Experimental Results
Experiments Experimental Results LIEML ranks 1 st in the most cases! The model of LE in LIEML is linear, but nonlinear in ML 2 , it is uneasy for LIEML to beat ML 2 with the less efficient linear way. ü The results of the experiment validate the effectiveness of our LIEML algorithm for multi-label learning.
Conclusion 4
Conclusion Major Contribution This paper proposes a novel multi-label learning method named LIEML, which enhances the labeling information by extending logical labels into numerical labels. The labeling information is enhanced by leveraging the underlying low- rank structure in the stacked matrix. More Information My personal website: Our lab website: https://antao.netlify.com/ http://palm.seu.edu.cn/
Thank You!
The LIEMLAlgorithm Label Enhancement To optimize the target function T 1 : Step 1: gradient decent Alternative Step 2: shrinkage ü To improve the speed of convergence, we begin with a large value µ 1 for µ , and decay as .
The LIEMLAlgorithm Predictive Model Induction To minimize T 2 ( Θ ; m ), we use an iterative quasi-Newton method called Iterative Re-Weighted Least Square (IRWLS). ≈ The quadratic problem can be solved as: • , , , , , . The solution for the next iteration ( Θ (k+1 ) and m ( k+1) ) of is obtained via a line search algorithm along ( Θ and m ) .
Experiments Experiment Configuration
Experiments Experiment Configuration Zhi-Hua Zhou, and Min-Ling Zhang. "Multi-label Learning." (2017): 875-881.
Introduction In traditional multi-label learning: Some features Model learning An instance predicting No Label 1 Yes No No Label 2 . . . . . . . . . Yes Yes Label d
Recommend
More recommend