Towards Fair Machine Learning Software: Understanding and Addressing Model Bias Through Counterfactual Thinking
Loading...
Links to Files
Author/Creator ORCID
Date
2023-02-16
Type of Work
Department
Program
Citation of Original Publication
Rights
This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
CC0 1.0 Universal (CC0 1.0) Public Domain Dedication
CC0 1.0 Universal (CC0 1.0) Public Domain Dedication
Subjects
Abstract
The increasing use of Machine Learning (ML) software can lead
to unfair and unethical decisions, thus fairness bugs in software
are becoming a growing concern. Addressing these fairness bugs
often involves sacrificing ML performance, such as accuracy. To
address this issue, we present a novel counterfactual approach that
uses counterfactual thinking to tackle the root causes of bias in ML
software. In addition, our approach combines models optimized for
both performance and fairness, resulting in an optimal solution in
both aspects. We conducted a thorough evaluation of our approach
on 10 benchmark tasks using a combination of 5 performance metrics, 3 fairness metrics, and 15 measurement scenarios, all applied
to 8 real-world datasets. The conducted extensive evaluations show
that the proposed method significantly improves the fairness of ML
software while maintaining competitive performance, outperforming state-of-the-art solutions in 84.6% of overall cases based on a
recent benchmarking tool.