Evaluation of Differential Privacy and Federated Learning for AI-Driven Customer Service Applications

Main Article Content

Yajing Zhang

Abstract

The proliferation of artificial intelligence in customer service applications has heightened concerns about the protection of sensitive consumer data. This paper presents a comprehensive comparative evaluation of privacy-preserving techniques, specifically differential privacy and federated learning, within AI-driven customer service contexts. A multidimensional evaluation framework is proposed to assess these techniques across security robustness, model accuracy, computational efficiency, and algorithmic fairness. An experimental analysis of customer interaction datasets reveals that federated learning achieves 94.2% accuracy retention while maintaining privacy guarantees, whereas differential privacy mechanisms offer superior protection against membership inference attacks at a 12.3% accuracy trade-off. The findings provide actionable recommendations for enterprises seeking to balance data protection compliance with service quality optimization, contributing to the development of trustworthy AI systems aligned with regulatory requirements.

Article Details

Section

Articles

How to Cite

Evaluation of Differential Privacy and Federated Learning for AI-Driven Customer Service Applications. (2026). Journal of Sustainability, Policy, and Practice, 2(2), 67-78. https://schoalrx.com/index.php/jspp/article/view/99

References

1. V. Mothukuri, R. M. Parizi, S. Pouriyeh, Y. Huang, A. Dehghantanha, and G. Srivastava, "A survey on security and privacy of federated learning," Future Generation Computer Systems, vol. 115, pp. 619-640, 2021. https://doi.org/10.1016/j.future.2020.10.007

2. P. Kairouz et al., "Advances and open problems in federated learning," Foundations and Trends in Machine Learning, vol. 14, no. 1-2, pp. 1-210, 2021. https://doi.org/10.1561/2200000083

3. T. Li, A. K. Sahu, A. Talwalkar, and V. Smith, "Federated learning: Challenges, methods, and future directions," IEEE Signal Processing Magazine, vol. 37, no. 3, pp. 50-60, 2020. https://doi.org/10.1109/MSP.2020.2975749

4. C. Dwork and A. Roth, "The algorithmic foundations of differential privacy," Foundations and Trends in Theoretical Computer Science, vol. 9, no. 3-4, pp. 211-407, 2014. https://doi.org/10.1561/0400000042

5. M. Abadi et al., "Deep learning with differential privacy," in Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, pp. 308-318, 2016. https://doi.org/10.1145/2976749.2978318

6. B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, "Communication-efficient learning of deep networks from decentralized data," in Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), pp. 1273-1282, 2017.

7. Q. Yang, Y. Liu, T. Chen, and Y. Tong, "Federated machine learning: Concept and applications," ACM Transactions on Intelligent Systems and Technology, vol. 10, no. 2, pp. 1-19, 2019. https://doi.org/10.1145/3298981

8. P. Radanliev, "AI ethics: Integrating transparency, fairness, and privacy in AI development," Applied Artificial Intelligence, vol. 39, no. 1, p. 2463722, 2025. https://doi.org/10.1080/08839514.2025.2463722

9. National Institute of Standards and Technology, "Guidelines for evaluating differential privacy guarantees (NIST Special Publication 800-226)," U.S. Department of Commerce, 2024. https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-226.pdf

10. D. Pessach and E. Shmueli, "A review on fairness in machine learning," ACM Computing Surveys, vol. 55, no. 3, pp. 1-44, 2023. https://doi.org/10.1145/3494672

11. N. Mehrabi, F. Morstatter, N. Saxena, K. Lerman, and A. Galstyan, "A survey on bias and fairness in machine learning," ACM Computing Surveys, vol. 54, no. 6, pp. 1-35, 2021. https://doi.org/10.1145/3457607

12. G. Xu, H. Li, S. Liu, K. Yang, and X. Lin, "VerifyNet: Secure and verifiable federated learning," IEEE Transactions on Information Forensics and Security, vol. 15, pp. 911-926, 2020. https://doi.org/10.1109/TIFS.2019.2929409

13. J. Zhao et al., "PVD-FL: A privacy-preserving and verifiable decentralized federated learning framework," IEEE Transactions on Information Forensics and Security, vol. 17, pp. 2059-2073, 2022. https://doi.org/10.1109/TIFS.2022.3176191

14. A. Fu et al., "VFL: A verifiable federated learning with privacy-preserving for big data in industrial IoT," IEEE Transactions on Industrial Informatics, vol. 18, no. 5, pp. 3316-3326, 2022. https://doi.org/10.1109/TII.2020.3036166

15. R. J. Chen et al., "Algorithmic fairness in artificial intelligence for medicine and healthcare," Nature Biomedical Engineering, vol. 7, no. 6, pp. 719-742, 2023. https://doi.org/10.1038/s41551-023-01056-8