Ethical AI Integration in Cybersecurity Operations: A Framework for Bias Mitigation and Human Oversight in Security Decision Systems
Keywords:
AI ethics, cybersecurity, algorithmic bias, human oversight, explainable AI, HITL, HOTL, ethical design, bias mitigation, security decision systemsAbstract
One of the emerging ethical issues regarding artificial intelligence (AI) is the use of AI in cybersecurity, particularly algorithmic fairness, transparency, and oversight. This article introduces a mitigated methodology for ethical AI integration, incorporating considerations of fairness, accountability, and human-centered design into security decision-making processes. Based on a synthesis of existing literature, technical case studies, and normative models, the paper presents the main oversight mechanisms, i.e., Human-in-the-Loop (HITL) and Human-on-the-Loop (HOTL) oversight, explainable AI interfaces, and continuous feedback units. The results not only illustrate the potential and the constraints of applying AI ethically in cybersecurity but also point out the most important directions future research and collaboration between various disciplines should take.
References
Brundage, M., Avin, S., Wang, J., Belfield, H., Krueger, G., Hadfield, G., Khlaaf, H., Yang, J., Toner, H., Fong, R., Maharaj, T., Koh, P. W., Hooker, S., Leung, J., Trask, A., Bluemke, E., Lebensold, J., O'Keefe, C., Koren, M., . . . Anderljung, M. (2020). Toward trustworthy AI Development: Mechanisms for supporting verifiable claims. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2004.07213
Cath, C. (2018). Governing artificial intelligence: ethical, legal and technical opportunities and challenges. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 376(2133), 20180080. https://doi.org/10.1098/rsta.2018.0080
Leslie, D. (2019). Understanding artificial intelligence ethics and safety: A guide for the responsible design and implementation of AI systems in the public sector. The Alan Turing Institute. https://doi.org/10.5281/zenodo.3240529
Ryan, M. (2018). Ethics of Public Use of AI and Big Data. ORBIT Journal. 2. 10.29297/orbit.v2i1.101.
Liz Rogers, IBM Security (2019). Bringing the Security Analyst into the Loop: From Human-Computer Interaction to Human-Computer Collaboration. EPIC Proceedings pp 341–361, ISSN 1559-8918, https://www.epicpeople.org/bringing-security-analyst-into-loop-human-computer-interaction-collaboration/
Mamman, H., Basri, S., Balogun, A., Imam, A. A., Kumar, G., & Capretz, L. F. (2024). Unbiasing on the fly: Explanation-Guided human oversight of machine learning system decisions. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2406.17906
Eastgate Software (September 13, 2024). AI in Cybersecurity: Key Case Studies and Breakthroughs. https://medium.com/%40eastgate/ai-in-cybersecurity-key-case-studies-and-breakthroughs-39bc72ce54ea
Brundage, Miles & Avin, Shahar & Clark, J. & Toner, H. & Eckersley, P. & Garfinkel, B. & Dafoe, A. & Scharre, P. & Zeitzoff, T. & Filar, B. & Roff, H. & Allen, G. & Steinhardt, J. & Flynn, C. & O Heigeartaigh, Sean & Beard, Sj & Belfield, Haydn & Farquhar, Sebastian & Amodei, Dario. (2018). The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation. 10.48550/arXiv.1802.07228.
Christiano, P. F., Leike, J., Brown, T., Martic, M., Legg, S., & Amodei, D. (2017). Deep reinforcement learning from human preferences. Advances in Neural Information Processing Systems, 30.
Cowls, J., & Floridi, L. (2019). A unified framework of five principles for AI in society. Harvard Data Science Review, 1(1). https://doi.org/10.1162/99608f92.8cd550d1
Doshi-Velez, F., & Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608.
https://doi.org/10.48550/arXiv.1702.08608
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). "Why Should I Trust You?": Explaining the Predictions of Any Classifier. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. https://doi.org/10.1145/2939672.2939778
Amershi Saleema, Dan Weld, Mihaela Vorvoreanu, Adam Fourney, Besmira Nushi, Penny Collisson, Jina Suh, Shamsi Iqbal, Paul N. Bennett, Kori Inkpen, Jaime Teevan, Ruth Kikin-Gil, & Eric Horvitz. (2019). Guidelines for Human-AI Interaction. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI'19' 19). Association for Computing Machinery, New York, NY, USA, Paper 3, 1–13. https://doi.org/10.1145/3290605.3300233
Barocas, S., Hardt, M., & Narayanan, A. (2023). Fairness and machine learning. http://fairmlbook.org/
Gunning, D., & Aha, D. (2019). DARPA's Explainable Artificial Intelligence (XAI) program. AI Magazine, 40(2), 44–58. https://doi.org/10.1609/aimag.v40i2.2850
Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., & Galstyan, A. (2021). A survey on bias and fairness in machine learning. ACM Computing Surveys, 54(6), 1–35. https://doi.org/10.1145/3457607
Cathy O'Neil. (2016). Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown Publishing Group, USA.
Raji, I. D., Smart, A., White, R. N., Mitchell, M., Gebru, T., Hutchinson, B., ... & Barnes, P. (2020). Closing the AI accountability gap: Defining an end-to-end framework for internal algorithmic auditing. Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 33–44. https://doi.org/10.1145/3351095.3372873
Sterz, S., Baum, K., Biewer, S., Hermanns, H., Lauber-Rönsberg, A., Meinel, P., & Langer, M. (2024, June). On the quest for effectiveness in human oversight: Interdisciplinary perspectives. In Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency (pp. 2495-2507).
Donald Farmer (December 27, 2024). TreeHive Strategy. Human oversight enables automated data governance. https://www.techtarget.com/searchdatamanagement/opinion/Human-oversight-enables-automated-data-governance
Binns Reuben (2018). Fairness in Machine Learning: Lessons from Political Philosophy. https://proceedings.mlr.press/v81/binns18a.html
Crawford, K. (2022). Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence. Perspectives on Science and Christian Faith. 74. 61–62. 10.56315/PSCF3-22Crawford.
Deeks, A., The Judicial Demand for Explainable Artificial Intelligence (August 1, 2019). 119 Colum. L. Rev. __ (2019 Forthcoming), Virginia Public Law and Legal Theory Research Paper No. 2019-51, Available at SSRN: https://ssrn.com/abstract=3440723
Mittelstadt, B. D., Allo, P., Taddeo, M., Wachter, S., & Floridi, L. (2016). The ethics of algorithms: Mapping the debate. Big Data & Society, 3(2). https://doi.org/10.1177/2053951716679679 (Original work published 2016)
Morley, J., Floridi, L., Kinsey, L., & Elhalal, A. (2019). From What to How: An Initial Review of Publicly Available AI Ethics Tools, Methods, and Research to Translate Principles into Practices. Science and Engineering Ethics, 26(4), 2141–2168. https://doi.org/10.1007/s11948-019-00165-5
Rahwan, I., Cebrian, M., Obradovich, N., Bongard, J., Bonnefon, J., Breazeal, C., Crandall, J. W., Christakis, N. A., Couzin, I. D., Jackson, M. O., Jennings, N. R., Kamar, E., Kloumann, I. M., Larochelle, H., Lazer, D., McElreath, R., Mislove, A., Parkes, D. C., Pentland, A. '., . . . Wellman, M. (2019). Machine behaviour. Nature, 568(7753), 477–486. https://doi.org/10.1038/s41586-019-1138-y
Inioluwa Deborah Raji & Joy Buolamwini. (2019). Actionable Auditing: Investigating the Impact of Publicly Naming Biased Performance Results of Commercial AI Products. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society (AIES' 19). Association for Computing Machinery, New York, NY, USA, 429–435. https://doi.org/10.1145/3306618.3314244
Sandra Wachter, Brent Mittelstadt, Luciano Floridi, Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation, International Data Privacy Law, Volume 7, Issue 2, May 2017, Pages 76–99, https://doi.org/10.1093/idpl/ipx005
Weller, A. (2019). Transparency: Motivations and Challenges. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L., Müller, KR. (eds) Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. Lecture Notes in Computer Science(), vol 11700. Springer, Cham. https://doi.org/10.1007/978-3-030-28954-6_2
Singhal A, Neveditsin N, Tanveer H, Mago V. Toward Fairness, Accountability, Transparency, and Ethics in AI for Social Media and Health Care: Scoping Review. JMIR Med Inform. 2024 April 3;12:e50048. doi: 10.2196/50048. PMID: 38568737; PMCID: PMC11024755.
Mokander, J., Morley, J., Taddeo, M. & Floridi, L. (2021). Ethics-Based Auditing of Automated Decision-Making Systems: Nature, Scope, and Limitations. 10.48550/arXiv.2110.10980.
Turner Nicol Lee, Paul Resnick, and Genie Barton (May 22, 2019). Algorithmic bias detection and mitigation: Best practices and policies to reduce consumer harms. https://www.brookings.edu/articles/algorithmic-bias-detection-and-mitigation-best-practices-and-policies-to-reduce-consumer-harms/
Wikipedia. Algorithmic accountability. https://en.wikipedia.org/wiki/Algorithmic_accountability?
Wikipedia. Joy Buolamwini. https://en.wikipedia.org/wiki/Joy_Buolamwini
https://redresscompliance.com/ethical-issues-ai-cybersecurity/
David Caswell, Sabthagiri Saravanan Chandramohan, Deborshi Dutt, Chris Knackstedt, Vikram Reddy Kunchala, David Mapgaonkar, Mike Morris, Abdul Rahman, Kate Fusillo Schmidt, Niels van de Vorle (2024). The CISO's Guide to Generative AI. https://www.deloitte.com/
Charmet, F., Tanuwidjaja, H.C., Ayoubi, S. et al. Explainable artificial intelligence for cybersecurity: a literature survey. Ann. Telecommun. 77, 789–812 (2022). https://doi.org/10.1007/s12243-022-00926-7
Akitra (September 16, 2024) Cybersecurity: Balancing Security Needs with Algorithmic Bias and Transparency. https://medium.com
Bruschi, D., Diomede, N. A framework for assessing AI ethics with cybersecurity applications. AI Ethics 3, 65–72 (2023). https://doi.org/10.1007/s43681-022-00162-8
Rjoub, G., Bentahar, J., Wahab, O. A., Mizouni, R., Song, A., Cohen, R., Otrok, H., & Mourad, A. (2023). A survey on Explainable Artificial intelligence for Cybersecurity. IEEE Transactions on Network and Service Management, 20(4), 5115–5140. https://doi.org/10.1109/tnsm.2023.3282740
Badi, Sadi. (2024). Ethical Implications of Integrating AI in Cybersecurity Systems: A Comprehensive Examination. International Journal of Applied Mathematics and Computer Science. 56–63.
Roman Panarin, Mekan Bairyev (May 2023) The Role of Artificial Intelligence in Cybersecurity. https://maddevs.io/blog/artificial-intelligence-in-cybersecurity/
Victoria Shutenko (08 August 2024) AI in Cybersecurity: Exploring the Top 6 Use Cases. https://www.techmagic.co/blog/ai-in-cybersecurity
Embedded Machine Intelligence Lab (Feb 20, 2024) Human-in-the-Loop Learning. https://ghasemzadeh.com/project/human-in-the-loop-learning/
Liz Ticong (April 29, 2024) AI in Cybersecurity: The Comprehensive Guide to Modern Security. https://www.datamation.com/security/ai-in-cybersecurity/
Prof. Norbert Pohlmann (October 2024) ARTIFICIAL INTELLIGENCE AND IT SECURITY - MORE SECURITY, MORE THREATS. https://www.dotmagazine.online/issues/digital-security-trust-consumer-protection/artificial-intelligence-it-security
Muniyandi, V. (2022). Harnessing Roslyn for advanced code analysis and optimization in cloud-based .NET applications on Microsoft Azure. International Journal of Communication Networks and Security, 14(4), 979-990.
Muniyandi, V. (2021). Extending Roslyn for custom code analysis and refactoring in large enterprise applications. International Journal of Science and Technology Research Archive, 3, 271-283.
Muniyandi, V. (2022). Harnessing Roslyn for advanced code analysis and optimization in cloud-based .NET applications on Microsoft Azure. International Journal of Communication Networks and Security, 14(4), 979-990.
Muniyandi, V. (2021). Extending Roslyn for custom code analysis and refactoring in large enterprise applications. International Journal of Science and Technology Research Archive, 3, 271-283.
Muniyandi, V. (2024). Design and Deployment of a Generative AI Copilot for Veterinary Practice Management Using Azure OpenAI and RAG Architecture. Available at SSRN 5342838.
Muniyandi, V. (2024). AI-Powered Document Processing with Azure Form Recognizer and Cognitive Search. Journal of Computational Analysis and Applications, 33(5).
Chellu, R. (2021). Secure Containerized Microservices Using PKI-Based Mutual TLS in Google Kubernetes Engine.
Chellu, R. (2022). Spectral Analysis of Cryptographic Hash Functions Using Fourier Techniques. Journal of Computational Analysis and Applications, 30(2).
Chellu, R. AI-Powered Intelligent Disaster Recovery and File Transfer Optimization for IBM Sterling and Connect: Direct in Cloud-Native Environments.
Chellu, R. (2024). Intelligent Data Movement: Leveraging AI to Optimize Managed File Transfer Performance Across Modern Enterprise Networks.
Chellu, R. Adaptive Quantum-Safe PKI Solutions for Nano-IoT Security Leveraging Cognitive Computing.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Well Testing Journal

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
This license requires that re-users give credit to the creator. It allows re-users to distribute, remix, adapt, and build upon the material in any medium or format, for noncommercial purposes only.