Multilabel classication is an extension of conventional classification in which a single instance can be associated with multiple labels. Recent research has shown that, just like for standard classication, instance-based learning algorithms relying on the nearest neighbor estimation
principle can be used quite successfully in this context. In this paper, we propose a new instance-based approach to multilabel classification, which is based on calibrated label ranking, a recently proposed
framework that unies multilabel classication and label ranking. Within
this framework, instance-based prediction is realized is the form of MAP
estimation, assuming a statistical distribution called the Mallows model.
The 7 Things I Know About Cyber Security After 25 Years | April 2024
Visitor Hotel Preference Learning Using Label Ranking
1. Weiwei Cheng & Eyke Hüllermeier
Knowledge Engineering & Bioinformatics Lab
Department of Mathematics and Computer Science
University of Marburg, Germany
2. Label Ranking (an example)
Learning visitor’s preferences on hotels
Golf ≻ Park ≻ Krim
label ranking
Krim ≻ Golf ≻ Park
visitor 1
Krim ≻ Park ≻ Golf
visitor 2
Park ≻ Golf ≻ Krim
visitor 3
visitor 4
new visitor ???
where the visitor could be described by feature vectors,
e.g., (gender, age, place of birth, is a professor, …)
1/15
3. Label Ranking (an example)
Learning visitor’s preferences on hotels
Golf Park Krim
visitor 1 1 2 3
visitor 2 2 3 1
visitor 3 3 2 1
visitor 4 2 1 3
new visitor ? ? ?
π(i) = position of the i-th label in the ranking
1: Golf 2: Park 3: Krim
2/15
4. Label Ranking (more formally)
Given:
a set of training instances
a set of labels
for each training instance : a set of pairwise preferences
of the form (for some of the labels)
Find:
A ranking function ( mapping) that maps each
to a ranking of (permutation ) and generalizes well
in terms of a loss function on rankings (e.g., Kendall’s tau)
3/15
6. Instance-based Label Ranking
nearest neighbor
Target function is estimated (on demand) in a local way.
Distribution of rankings is (approx.) constant in a local region.
Core part is to estimate the locally constant model.
5/15
7. Probabilistic Model for Ranking
Mallows model (Mallows, Biometrika, 1957)
with
center ranking
spread parameter
and is a metric on permutations
6/15
8. Inference
Observation Extensions
An example:
7/15
9. Inference (Cont.)
For observations :
Maximum Likelihood Estimation (MLE) becomes a difficult,
non-convex optimization problem!
8/15
10. Special Structure of Observations
In the multilabel classification context, the observation is
a special type of partial ranking. It contains:
two tie groups (i.e. relevant & irrelevant label set)
all labels (i.e. no label is missing)
By exploiting this special structure , MLE can be made
much more efficiently!
9/15
11. First Theorem
Sorting the labels according to their
frequency of occurrence in the neighborhood
Problem: What about ties?
Solution: Replacing MLE with Bayes estimation.
10/15
12. Second Theorem
A simple prediction procedure:
1. Sort labels with their frequency in the neighborhood
2. Break ties with frequency outside
11/15
16. Our Contributions
A new instance-based multilabel classifier with state-
of-the-art predictive accuracy;
It is computationally very efficient;
with very simple prediction procedure, justified by an
underlying probabilistic ranking model.
The End