Partial Multi-Label Learning via Label-Specific Feature Corrections

Frontier Research in Partial Multi-Label Learning: A New Method Based on Label-Specific Feature Corrections

In recent years, partial multi-label learning (PML) has become a hot research topic in the field of machine learning. With the rise of crowdsourcing platforms, the cost of data annotation has dropped significantly, but the quality of annotations has also decreased—resulting in candidate label sets inevitably containing irrelevant labels. These noisy labels not only increase the difficulty of learning tasks but may also mislead model performance. Therefore, studying how to efficiently learn from noisy data has become a critical issue in academic research. This article focuses on a study that provides significant insights into this field, proposing a new method called PASE (Partial Multi-Label Learning via Label-Specific Feature Corrections), which offers an innovative solution distinct from traditional “disambiguation” strategies.


Background and Problem Definition

Partial multi-label learning (PML) is a typical weakly-supervised learning problem. In this setting, an instance is usually assigned a set of candidate labels, only a subset of which are actual relevant labels, while others belong to irrelevant categories. This type of problem has broad applications in real-world scenarios, such as image classification, semantic analysis, and more. However, the presence of noisy labels renders traditional multi-label learning methods difficult to adapt directly to the PML context.

To date, most methods have focused on a “disambiguation” strategy, which attempts to recover the ground-truth labels from the candidate label set by designing various assumptions and algorithms before training the learning model. Although intuitive, this approach faces significant limitations, such as relying on prior assumptions about data distribution that are often invalid in real-world scenarios, leading to error accumulation in predictive models. This study raises an intriguing question: In PML tasks, is it possible to bypass label-space manipulation and directly solve the problem by adjusting instance features?


Paper Source and Author Information

This paper, titled Partial Multi-Label Learning via Label-Specific Feature Corrections, was co-authored by Jun-Yi Hang and Min-Ling Zhang from Southeast University. This study is affiliated with the Key Laboratory of Computer Network and Information Integration (Ministry of Education) at Southeast University and was published in the Science China Information Sciences journal, March 2025 (Volume 68, Issue 3). The proposed PASE method provides a novel solution that moves away from traditional label-disambiguation strategies.


Research Methods and Workflow

The core idea of this study is to eliminate potential ambiguities in PML data by adjusting the instance feature space, thereby directly building a predictive model under a meta-learning framework. Below, we detail the research methodology, including the workflow, experimental design, and data analysis.

Core Workflow

  1. Problem Modeling and Feature Correction Function Design
    In PASE, the researchers formalized the PML problem as an alternating optimization process between feature correction and model learning. The feature correction process is designed as a label-specific mapping function to adjust instance features based on each category’s characteristics. This adjustment aims to relocate instances to the correct position in the feature space, aligning them with the desired prediction model’s targets.

The correction function is specifically defined as a combination of scaling and translation:

   gk(eφ(x);ψk) = wk(eφ(x);ψk)·eφ(x) + bk(eφ(x);ψk)

where wk(·) and bk(·) are small hyper-networks generating parameters, eφ(x) represents the embedding of instance features, and the correction parameters ψk are designed specifically for each label.

  1. Meta-Learning Framework and Bi-Level Optimization
    PASE formulates the learning process as a bi-level optimization problem: The outer optimization minimizes errors on a clean validation set, indirectly improving model generalization by adjusting feature correction parameters (ψ). The inner optimization focuses on minimizing the loss on corrected training data. The alternating optimization process updates the feature correction and prediction model parameters iteratively, gradually converging to a globally better solution.

  2. Algorithm Implementation and Solution
    To improve computational efficiency, the researchers decomposed the bi-level optimization problem into a series of solvable subproblems and implemented the PASE algorithm using the PyTorch deep learning framework. During optimization, the Adam optimizer and automatic differentiation techniques were employed to accelerate gradient computation. The workflow’s pseudocode is clearly presented in the paper, ensuring reproducibility.


Experiment Design and Data Analysis

Experiment Setup

The researchers conducted comprehensive experiments and comparative studies on 21 benchmark datasets, including five real-world PML datasets (e.g., YeastBP, Music Emotion) and six synthetic PML datasets derived from multi-label data mining (e.g., Corel16K-S1, Delicious). The datasets span diverse domains such as biology, text, images, and video, with label counts ranging from 10 to 200.

To ensure fairness, the datasets were split, with 90% used for training and validation and the remaining 10% for testing.

Comparative Results

PASE demonstrated significantly superior performance across multiple evaluation metrics, including Average Precision, Ranking Loss, Coverage, and Hamming Loss. Key observations include:

  • PASE achieved the best performance in 79% of cases across the 21 datasets.
  • Compared to traditional disambiguation-based algorithms like PML-MD, PASE significantly reduced both ranking loss and coverage, highlighting its effectiveness in handling high-noise label distributions.

For instance, PASE achieved an average precision of 0.362 on the YeastBP dataset, compared to only 0.284 for the traditional FPML algorithm. On the MediaMill dataset, ranking loss was reduced from 0.140 to 0.121.

Visualization Verification

On a synthetic dataset with a dual Gaussian distribution, the researchers visually demonstrated the effectiveness of PASE’s feature correction. After correction, irrelevant “false positive” points were successfully relocated to the correct category regions, and the corresponding decision boundaries aligned more closely with the ground-truth label distribution.


Research Conclusions and Significance

Conclusions:

PASE challenges the dominance of traditional “label disambiguation” strategies by introducing label-specific feature correction as a novel solution to the PML problem. In experiments, PASE not only outperformed traditional methods significantly on various evaluation metrics but also exhibited strong robustness, particularly in high-noise data scenarios.

Scientific and Practical Value:

  • Scientific Contribution: This is the first time a feature correction strategy has been proposed to solve the PML problem, providing a new perspective for future research.
  • Practical Applications: The method is suitable for noisy data learning tasks across various domains, such as automatic image annotation and text multi-label classification.

Research Highlights:

  1. The creative design of a label-specific correction method enables the modeling of unique information for each category.
  2. Incorporating a meta-learning framework uses feedback information from validation sets as an optimization strategy, significantly reducing overfitting risks.
  3. Comprehensive and systematic experimental validation across multiple domains demonstrates strong generalizability.

Future Directions

At the conclusion of the paper, the authors state that label-specific feature correction may lead to a class imbalance issue, which merits deeper exploration in future research. Additionally, incorporating balancing methods from other domains may further enhance PASE’s performance in high-noise scenarios.

This study points to a new direction for research in the field of partial multi-label learning, deepening our understanding of feature modeling in weakly-supervised learning scenarios. The introduction of the PASE method is expected to receive broader attention and applications in the future.