When Biased Humans Meet Debiased AI: A Case Study in College Major Recommendation
dc.contributor.author | Wang, Clarice | |
dc.contributor.author | Wang, Kathryn | |
dc.contributor.author | Bian, Andrew Y. | |
dc.contributor.author | Islam, Rashidul | |
dc.contributor.author | Keya, Kamrun Naher | |
dc.contributor.author | Foulds, James | |
dc.contributor.author | Pan, Shimei | |
dc.date.accessioned | 2023-08-21T22:39:16Z | |
dc.date.available | 2023-08-21T22:39:16Z | |
dc.date.issued | 2023-08-01 | |
dc.description.abstract | Currently, there is a surge of interest in fair Artificial Intelligence (AI) and Machine Learning (ML) research which aims to mitigate discriminatory bias in AI algorithms, e.g. along lines of gender, age, and race. While most research in this domain focuses on developing fair AI algorithms, in this work, we examine the challenges which arise when humans and fair AI interact. Our results show that due to an apparent conflict between human preferences and fairness, a fair AI algorithm on its own may be insufficient to achieve its intended results in the real world. Using college major recommendation as a case study, we build a fair AI recommender by employing gender debiasing machine learning techniques. Our offline evaluation showed that the debiased recommender makes fairer career recommendations without sacrificing its accuracy in prediction. Nevertheless, an online user study of more than 200 college students revealed that participants on average prefer the original biased system over the debiased system. Specifically, we found that perceived gender disparity is a determining factor for the acceptance of a recommendation. In other words, we cannot fully address the gender bias issue in AI recommendations without addressing the gender bias in humans. We conducted a follow-up survey to gain additional insights into the effectiveness of various design options that can help participants to overcome their own biases. Our results suggest that making fair AI explainable is crucial for increasing its adoption in the real world. | en_US |
dc.description.sponsorship | This work was performed under the following inancial assistance award: 60NANB18D227 from U.S. Department of Commerce, National Institute of Standards and Technology. This material is based upon work supported by the National Science Foundation under Grant No.’s IIS2046381; IIS1850023; IIS1927486. Any opinions, indings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily relect the views of the National Science Foundation. | en_US |
dc.description.uri | https://dl.acm.org/doi/10.1145/3611313 | en_US |
dc.format.extent | 28 pages | en_US |
dc.genre | journal articles | en_US |
dc.genre | postprints | en_US |
dc.identifier | doi:10.13016/m2fixa-dyza | |
dc.identifier.citation | Wang, Clarice, Kathryn Wang, Andrew Y. Bian, Rashidul Islam, Kamrun Naher Keya, James Foulds, and Shimei Pan. “When Biased Humans Meet Debiased AI: A Case Study in College Major Recommendation.” ACM Transactions on Interactive Intelligent Systems, August 1, 2023. https://doi.org/10.1145/3611313. | en_US |
dc.identifier.uri | https://doi.org/10.1145/3611313 | |
dc.identifier.uri | http://hdl.handle.net/11603/29309 | |
dc.language.iso | en_US | en_US |
dc.publisher | ACM | en_US |
dc.relation.isAvailableAt | The University of Maryland, Baltimore County (UMBC) | |
dc.relation.ispartof | UMBC Information Systems Department Collection | |
dc.relation.ispartof | UMBC Faculty Collection | |
dc.relation.ispartof | UMBC Student Collection | |
dc.rights | Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org. | en_US |
dc.title | When Biased Humans Meet Debiased AI: A Case Study in College Major Recommendation | en_US |
dc.type | Text | en_US |
dcterms.creator | https://orcid.org/0000-0001-5276-5708 | en_US |
dcterms.creator | https://orcid.org/0000-0003-0935-4182 | en_US |
dcterms.creator | https://orcid.org/0000-0002-5989-8543 | en_US |