Multilabel classification has lately gained growing interest in the research community.
We implemented several methods, which make use of the standardized mlr framework. Every available binary learner can be used for multilabel problem transformation methods.
So if you’re interested in using several multilabel algorithms and want to know how to use them in the mlr framework, then this post is for you!
1) Introduction to multilabel classification
First, let me introduce you to multilabel classification. This is a classification problem, where every instance can have more than one label. Let’s have a look at a typical multilabel dataset (which I, of course, download from the OpenML server):
Here I took the scene dataset, where the features represent color information of pictures and the targets could be objects like beach, sunset, and so on.
As you can see above, one defining property of a multilabel dataset is, that the target variables (which are called labels) are binary. If you want to use your own data set, make sure to encode these variables in logical, where TRUE indicates the relevance of a label.
The basic idea behind many multilabel classification algorithms is to make use of possible correlation between labels. Maybe a learner is very good at predicting label 1, but rather bad at predicting label 2. If label 1 and label 2 are highly correlated, it may be beneficial to predict label 1 first and use this prediction as a feature for predicting label 2.
This approach is the main concept behind the so called problem transformation methods. The multilabel problem is transformed into binary classification problems, one for each label. Predicted labels are used as features for predicting other labels.
We implemented the following problem transformation methods:
- Classifier chains
- Nested stacking
- Dependent binary relevance
2) Let’s Train and Predict!
First we need to create a multilabel task.
We set a seed, because the classifier chain wrapper uses a random chain order.
Next, we train a learner. I chose the classifier chain approach together with a decision tree for the binary classification problems.
Now let’s train and predict on our dataset:
We also implemented common multilabel performance measures. Here is a list with available multilabel performance measures:
Here is how the classifier chains method performed:
3) Comparison Binary Relevance vs. Classifier Chains
Now let’s see if it can be beneficial to use predicted labels as features for other labels. Let us compare the performance of the classifier chains method with the binary relevance method (this method does not use predicted labels as features).
As can be seen here, it could indeed make sense to use more elaborate methods for multilabel classification, since classifier chains beat the binary relevance methods in all of these measures (Note, that hamming loss and subset01 are loss measures!).
Here I’ll show you how to use resampling methods in the multilabel setting. Resampling methods are key for assessing the performance of a learning algorithm. To read more about resampling, see the page on our tutorial.
First, we need to define a resampling strategy. I chose subsampling, which is also called Monte-Carlo cross-validation. The dataset is split into training and test set at a predefined ratio. The learner is trained on the training set, the performance is evaluated with the test set. This whole process is repeated many times and the performance values are averaged. In mlr this is done the following way:
Now we can choose a measure, which shall be resampled. All there is left to do is to run the resampling:
If you followed the mlr tutorial or if you are already familiar with mlr, you most likely saw, that using resampling in the multilabel setting isn’t any different than generally using resampling in mlr.
Many methods, which are available in mlr, like preprocessing, tuning or benchmark experiments can also be used for multilabel datasets and the good thing here is: the syntax stays the same!