Principle:Cleanlab Cleanlab Datalab Multilabel Label Issue Detection
| Knowledge Sources | |
|---|---|
| Domains | Data Quality, Multilabel Classification |
| Last Updated | 2026-02-09 00:00 GMT |
Overview
Multilabel label issue detection identifies examples in a multilabel classification dataset whose assigned label sets are likely incorrect, using model-predicted class probabilities to estimate per-example label quality.
Description
In multilabel classification, each example is associated with a set of labels rather than a single label. This makes annotation errors both more common and harder to detect, because an error can involve any combination of missing labels, extraneous labels, or substituted labels within the set.
Multilabel label issue detection addresses this by leveraging a trained model's predicted probabilities to assess whether the given label set for each example is consistent with the model's learned patterns. When the model strongly disagrees with the annotations, the example is flagged as a potential label issue.
This approach is critical for tasks such as:
- Image tagging: where images may be tagged with incorrect or incomplete sets of descriptors.
- Document categorization: where documents may be assigned to wrong topic categories.
- Medical diagnosis: where patient records may have missing or incorrect diagnostic codes.
The key insight is that multilabel errors are decomposable: the problem can be treated as multiple binary classification problems (one per class), and the per-class error estimates can be aggregated into a single quality score for the full label set.
Usage
Apply multilabel label issue detection when:
- Your dataset uses multilabel annotations and you suspect annotation errors.
- You have access to predicted class probabilities from a trained multilabel classifier.
- You want to prioritize examples for manual review based on likelihood of annotation error.
- You need to estimate the overall label quality of a multilabel dataset.
Theoretical Basis
Multilabel label issue detection in cleanlab is built on the following principles:
1. Per-class binary decomposition:
Each class k is treated as an independent binary classification problem. For a given example with predicted probability p_k for class k:
- If
p_k > 0.5, the model predicts the example belongs to classk. - If
p_k <= 0.5, the model predicts the example does not belong to classk.
2. Label quality scoring:
A per-example label quality score is computed using methods such as self-confidence, which measures how likely the given label set is according to the model. Higher scores indicate greater agreement between the model and the annotation; lower scores indicate potential errors. The score aggregates information across all classes for each example.
3. Issue identification:
Examples are flagged as label issues using confident learning principles adapted for the multilabel setting. The find_label_issues() function identifies examples where the model's confident predictions disagree with the given labels, using configurable filtering methods and noise fraction estimates.
4. Summary statistic:
The overall dataset quality is summarized as the mean label quality score across all examples:
dataset_score = (1/n) * sum(score_i for i in 1..n)
A lower mean score indicates a dataset with more label quality problems overall.
5. Predicted label derivation:
Predicted labels are obtained by thresholding the predicted probabilities at 0.5 and converting the resulting binary vector into an integer list representation using onehot2int(). This allows direct comparison between given and predicted label sets.