IBM J. Res. Dev. | 2019

AI Fairness 360: An extensible toolkit for detecting and mitigating algorithmic bias

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Abstract


Fairness is an increasingly important concern as machine learning models are used to support decision making in high-stakes applications such as mortgage lending, hiring, and prison sentencing. This article introduces a new open-source Python toolkit for algorithmic fairness, AI Fairness 360 (AIF360), released under an Apache v2.0 license (\n https://github.com/ibm/aif360 \n). The main objectives of this toolkit are to help facilitate the transition of fairness research algorithms for use in an industrial setting and to provide a common framework for fairness researchers to share and evaluate algorithms. The package includes a comprehensive set of fairness metrics for datasets and models, explanations for these metrics, and algorithms to mitigate bias in datasets and models. It also includes an interactive Web experience that provides a gentle introduction to the concepts and capabilities for line-of-business users, researchers, and developers to extend the toolkit with their new algorithms and improvements and to use it for performance benchmarking. A built-in testing infrastructure maintains code quality.

Volume 63
Pages 4:1-4:15
DOI 10.1147/jrd.2019.2942287
Language English
Journal IBM J. Res. Dev.

Full Text