arXiv Analytics

Sign in

arXiv:2401.16594 [cs.LG]AbstractReferencesReviewsResources

Consistent algorithms for multi-label classification with macro-at-$k$ metrics

Erik Schultheis, Wojciech Kotłowski, Marek Wydmuch, Rohit Babbar, Strom Borman, Krzysztof Dembczyński

Published 2024-01-29, updated 2024-06-20Version 2

We consider the optimization of complex performance metrics in multi-label classification under the population utility framework. We mainly focus on metrics linearly decomposable into a sum of binary classification utilities applied separately to each label with an additional requirement of exactly $k$ labels predicted for each instance. These "macro-at-$k$" metrics possess desired properties for extreme classification problems with long tail labels. Unfortunately, the at-$k$ constraint couples the otherwise independent binary classification tasks, leading to a much more challenging optimization problem than standard macro-averages. We provide a statistical framework to study this problem, prove the existence and the form of the optimal classifier, and propose a statistically consistent and practical learning algorithm based on the Frank-Wolfe method. Interestingly, our main results concern even more general metrics being non-linear functions of label-wise confusion matrices. Empirical results provide evidence for the competitive performance of the proposed approach.

Comments: This is the authors' version of the work accepted to ICLR 2024; the final version of the paper, errors and typos corrected, and minor modifications to improve clarity
Categories: cs.LG
Related articles: Most relevant | Search more
arXiv:1802.01697 [cs.LG] (Published 2018-02-05)
Deep Learning with a Rethinking Structure for Multi-label Classification
arXiv:2009.08607 [cs.LG] (Published 2020-09-18)
Compact Learning for Multi-Label Classification
arXiv:1602.07464 [cs.LG] (Published 2016-02-24)
Feature ranking for multi-label classification using Markov Networks