Fairness-Aware Credit Evaluation: Bias Detection and Mitigation Techniques for Inclusive Lending Practices

Main Article Content

Zhi Luo

Abstract

The widespread adoption of machine learning in credit assessment has raised concerns about algorithmic fairness and discriminatory outcomes affecting protected demographic groups. This paper investigates bias detection methodologies and mitigation techniques designed to promote equitable lending while maintaining predictive accuracy. We analyze three intervention categories: pre-processing data transformations, in-processing algorithmic constraints, and post-processing decision adjustments. Through empirical evaluation on credit datasets, we examine accuracy-fairness trade-offs and assess practical viability for financial institutions under regulatory compliance frameworks. Our comparative analysis demonstrates that mitigation strategies exhibit distinct performance characteristics depending on fairness metrics and business objectives. The findings provide guidance for practitioners balancing the precision of risk assessment with equitable treatment across diverse applicant populations.

Article Details

Section

Articles

How to Cite

Fairness-Aware Credit Evaluation: Bias Detection and Mitigation Techniques for Inclusive Lending Practices. (2026). Journal of Sustainability, Policy, and Practice, 2(2), 78-89. https://schoalrx.com/index.php/jspp/article/view/102

References

1. N. Kozodoi, J. Jacob, and S. Lessmann, "Fairness in credit scoring: Assessment, implementation and profit implications," European Journal of Operational Research, vol. 297, no. 3, pp. 1083-1094, 2022.

2. Y. Zhang, M. Chen, S. Liu, B. Zhang, and J. Tang, "Algorithmic fairness in financial services: A systematic literature review," Finance Research Letters, vol. 68, p. 105891, 2024.

3. C. Hurlin, C. Pérignon, and S. Saurin, "The fairness of credit scoring models," Management Science, vol. 72, no. 1, pp. 406-425, 2024.

4. P. B. de Laat, "Algorithmic discrimination in the credit domain: What do we know about it?" AI & Society, vol. 38, no. 6, pp. 2581-2601, 2023.

5. A. Fuster, P. Goldsmith-Pinkham, T. Ramadorai, and A. Walther, "Predictably unequal? The effects of machine learning on credit markets," Journal of Finance, vol. 77, no. 1, pp. 5-47, 2022.

6. A. Castelnovo, R. Crupi, G. Greco, D. Regoli, I. G. Penco, and A. C. Cosentini, "A clarification of the nuances in the fairness metrics landscape," Scientific Reports, vol. 12, no. 1, p. 4209, 2022.

7. R. K. Bellamy, K. Dey, M. Hind, S. C. Hoffman, S. Houde, K. Kannan, P. Lohia, J. Martino, S. Mehta, A. Mojsilovic, S. Nagar, K. N. Ramamurthy, J. Richards, D. Saha, P. Sattigeri, M. Singh, K. R. Varshney, and Y. Zhang, "AI Fairness 360: An extensible toolkit for detecting and mitigating algorithmic bias," IBM Journal of Research and Development, vol. 63, no. 4/5, pp. 4:1-4:15, 2019.

8. World Bank Group, "The use of alternative data in credit risk assessment: Opportunities, risks, and challenges," International Committee on Credit Reporting, 2023.

9. S. Agarwal, S. Alok, P. Ghosh, and S. Gupta, "Financial inclusion and alternate credit scoring for the millennials: Role of big data and machine learning in fintech," Working Paper, Indian School of Business, 2020.

10. R. L. Oaxaca, "Male-Female Wage Differentials in Urban Labor Markets," International Economic Review, vol. 14, no. 3, pp. 693--709, 1973.

11. M. Hardt, E. Price, and N. Srebro, "Equality of Opportunity in Supervised Learning," Advances in Neural Information Processing Systems (NeurIPS), vol. 29, pp. 3315--3323, 2016.

12. A. Chouldechova, "Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments," Big Data, vol. 5, no. 2, pp. 153--163, 2017.

13. M. Feldman, S. A. Friedler, J. Moeller, C. Scheidegger, and S. Venkatasubramanian, "Certifying and Removing Disparate Impact," in *Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining*, 2015, pp. 259--268.

14. E. Richardson, J. A. Weiss, and M. T. Schmid, "The receiver operating characteristic curve accurately compares treatment effect prediction models for binary outcomes," Patterns, vol. 5, no. 8, 2024.

15. B. H. Zhang, B. Lemoine, and M. Mitchell, "Mitigating Unwanted Biases with Adversarial Learning," in Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, 2018, pp. 335--340.