arXiv Analytics

Sign in

arXiv:2209.07850 [cs.LG]AbstractReferencesReviewsResources

FairGBM: Gradient Boosting with Fairness Constraints

André F Cruz, Catarina Belém, João Bravo, Pedro Saleiro, Pedro Bizarro

Published 2022-09-16Version 1

Machine Learning (ML) algorithms based on gradient boosted decision trees (GBDT) are still favored on many tabular data tasks across various mission critical applications, from healthcare to finance. However, GBDT algorithms are not free of the risk of bias and discriminatory decision-making. Despite GBDT's popularity and the rapid pace of research in fair ML, existing in-processing fair ML methods are either inapplicable to GBDT, incur in significant train time overhead, or are inadequate for problems with high class imbalance. We present FairGBM, a learning framework for training GBDT under fairness constraints with little to no impact on predictive performance when compared to unconstrained LightGBM. Since common fairness metrics are non-differentiable, we employ a ``proxy-Lagrangian'' formulation using smooth convex error rate proxies to enable gradient-based optimization. Additionally, our open-source implementation shows an order of magnitude speedup in training time when compared with related work, a pivotal aspect to foster the widespread adoption of FairGBM by real-world practitioners.

Related articles: Most relevant | Search more
arXiv:1806.06055 [cs.LG] (Published 2018-06-15)
Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees
arXiv:2304.04091 [cs.LG] (Published 2023-04-08)
Best Arm Identification with Fairness Constraints on Subpopulations
arXiv:2411.05318 [cs.LG] (Published 2024-11-08)
Fairness in Monotone $k$-submodular Maximization: Algorithms and Applications