How Explainable AI Reduces Bias

Authors

  • Rajarshi Roy, Sridharan Narayanan

Keywords:

interpretability, trustworthiness, AI

Abstract


This research paper investigates the role of Explainable Artificial Intelligence (XAI) in reducing bias in AI systems. As AI becomes increasingly prevalent in decision-making processes across various domains, concerns about algorithmic bias have grown. This study explores how XAI techniques can be leveraged to identify, mitigate, and prevent bias in AI models. Through a comprehensive analysis of existing literature, implementation of XAI models, and evaluation of their effectiveness in bias reduction, this research contributes to the ongoing efforts to develop more fair and transparent AI systems. The findings demonstrate that XAI techniques, when properly applied, can significantly reduce bias in AI models while improving their interpretability and trustworthiness.

Downloads

Download data is not yet available.

References

Adadi, A., & Berrada, M. (2018). Peeking inside the black-box: A survey on Explainable Artificial Intelligence (XAI). IEEE Access, 6, 52138-52160. https://ieeexplore.ieee.org/document/8466590

Amini, A., Soleimany, A. P., Schwarting, W., Bhatia, S. N., & Rus, D. (2019). Uncovering and mitigating algorithmic bias through learned latent structure. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society (pp. 289-295). https://dl.acm.org/doi/10.1145/3306618.3314243

Buolamwini, J., & Gebru, T. (2018). Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency (pp. 77-91). PMLR. http://proceedings.mlr.press/v81/buolamwini18a.html

Dodge, J., Liao, Q. V., Zhang, Y., Bellamy, R. K., & Dugan, C. (2019). Explaining models: an empirical study of how explanations impact fairness judgment. In Proceedings of the 24th International Conference on Intelligent User Interfaces (pp. 275-285). https://dl.acm.org/doi/10.1145/3301275.3302310

Gunning, D., & Aha, D. W. (2019). DARPA's explainable artificial intelligence program. AI Magazine, 40(2), 44-58. https://ojs.aaai.org/index.php/aimagazine/article/view/2850

Lundberg, S. M., & Lee, S. I. (2017). A unified approach to interpreting model predictions. In Advances in neural information processing systems (pp. 4765-4774). https://proceedings.neurips.cc/paper/2017/hash/8a20a8621978632d76c43dfd28b67767-Abstract.html

Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., & Galstyan, A. (2021). A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR), 54(6), 1-35. https://dl.acm.org/doi/10.1145/3457607

Ntoutsi, E., Fafalios, P., Gadiraju, U., Iosifidis, V., Nejdl, W., Vidal, M. E., ... & Staab, S. (2020). Bias in data‐driven artificial intelligence systems—An introductory survey. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 10(3), e1356. https://wires.onlinelibrary.wiley.com/doi/full/10.1002/widm.1356

Obermeyer, Z., Powers, B., Vogeli, C., & Mullainathan, S. (2019). Dissecting racial bias in an algorithm used to manage the health of populations. Science, 366(6464), 447-453. https://www.science.org/doi/10.1126/science.aax2342

Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. (2017). Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision (pp. 618-626). https://openaccess.thecvf.com/content_iccv_2017/html/Selvaraju_Grad-CAM_Visual_Explanations_ICCV_2017_paper.html

Downloads

Published

26.09.2021

How to Cite

Rajarshi Roy. (2021). How Explainable AI Reduces Bias. International Journal of Intelligent Systems and Applications in Engineering, 9(3), 136 –143. Retrieved from https://www.ijisae.org/index.php/IJISAE/article/view/6902

Issue

Section

Research Article