Home / Papers / Security in Machine Learning (ML) Workflows

Security in Machine Learning (ML) Workflows

88 Citations2024
Dinesh Reddy Chittibala, Srujan Reddy Jabbireddy
International Journal of Computing and Engineering

This paper proposes a holistic suite of strategies designed to enhance the security of ML workflows, including advanced data protection techniques like anonymization and encryption, model security enhancements through adversarial training and hardening, and the fortification of infrastructure security via secure computing environments and continuous monitoring.

Abstract

Purpose: This paper addresses the comprehensive security challenges inherent in the lifecycle of machine learning (ML) systems, including data collection, processing, model training, evaluation, and deployment. The imperative for robust security mechanisms within ML workflows has become increasingly paramount in the rapidly advancing field of ML, as these challenges encompass data privacy breaches, unauthorized access, model theft, adversarial attacks, and vulnerabilities within the computational infrastructure. Methodology: To counteract these threats, we propose a holistic suite of strategies designed to enhance the security of ML workflows. These strategies include advanced data protection techniques like anonymization and encryption, model security enhancements through adversarial training and hardening, and the fortification of infrastructure security via secure computing environments and continuous monitoring. Findings: The multifaceted nature of security challenges in ML workflows poses significant risks to the confidentiality, integrity, and availability of ML systems, potentially leading to severe consequences such as financial loss, erosion of trust, and misuse of sensitive information. Unique Contribution to Theory, Policy and Practice: Additionally, this paper advocates for the integration of legal and ethical considerations into a proactive and layered security approach, aiming to mitigate the risks associated with ML workflows effectively. By implementing these comprehensive security measures, stakeholders can significantly reinforce the trustworthiness and efficacy of ML applications across sensitive and critical sectors, ensuring their resilience against an evolving landscape of threats.