Are Your AI Models at Risk of Security Flaws and Data Exfiltration?

When Palo Alto Networks’ Unit 42 recently discovered two significant security vulnerabilities in Google’s Vertex AI machine learning (ML) platform, it highlighted notable risks. These vulnerabilities have the potential to lead to unauthorized access, privilege escalation, and critical data exfiltration if exploited. Vertex AI, introduced in May 2021, offers a scalable environment for training and deploying custom ML models and AI applications, making this discovery particularly significant given the widespread reliance on these platforms for advanced AI solutions.

Vertex AI Pipelines’ Custom Job Permissions

Exploiting Custom Job Permissions

The first identified flaw concerns the Vertex AI Pipelines feature, specifically its handling of custom job permissions. Vertex AI Pipelines automate and monitor MLOps workflows, which are essential for efficient machine learning operations. However, the exploitation of these custom job permissions can allow threat actors to create a custom job that executes a specially-crafted image. This image, in turn, can initiate a reverse shell, granting the attacker backdoor access to the system. This method enables privilege escalation, providing attackers with control over extensive permissions, including managing storage buckets and accessing BigQuery tables.

With these elevated permissions, attackers have the potential to misuse the access to internal Google Cloud repositories. This means they can download sensitive and proprietary data, thereby undermining the confidentiality and integrity of machine learning work conducted within the Vertex AI platform. By gaining control over storage buckets and accessing critical data from BigQuery tables, the attackers can easily extract information that is highly valuable and potentially damaging if exposed. Consequently, this flaw presents a concerning threat, particularly given the significant data managed by organizations leveraging Vertex AI for their ML operations.

Risks of Privilege Escalation

Privilege escalation through Vertex AI Pipelines’ custom job permissions represents a severe risk for organizations using the platform. By manipulating the permissions allowed for custom jobs, attackers can effectively extend their reach within the cloud environment. This not only jeopardizes the specific projects being targeted but also exposes adjacent areas of the cloud infrastructure to potential exploits. The interconnected nature of cloud permissions means that successful privilege escalation could lead to broader-scale data breaches and infrastructure compromise. This highlights the necessity for stringent control mechanisms and permission audits to safeguard against such exploitations.

Poisoned Model Deployment in Tenant Projects

Deployment of Manipulated Models

The second vulnerability involves the deployment of a manipulated, or "poisoned," model within a tenant project. This particular vulnerability stands out because once the poisoned model is deployed, it initiates a reverse shell, enabling the attacker to exploit the read-only permissions of the "custom-online-prediction" service account. This access permits the enumeration of Kubernetes clusters and the retrieval of credentials. The attacker can then execute arbitrary kubectl commands, further enhancing the exploitation potential. Such lateral movement between the Google Cloud Platform (GCP) and the Google Kubernetes Engine (GKE) is facilitated by linked permissions through IAM Workload Identity Federation.

Attackers can use this access to extract container images using tools like crictl and capture authentication tokens. These tokens enable them to view and export large-language models (LLMs) and their fine-tuned adapters, posing substantial data exfiltration risks. The capacity to exfiltrate proprietary models and data from an organization’s cloud infrastructure highlights the critical need for vigilant monitoring and defense strategies to prevent unauthorized deployments and access.

Implications of Lateral Movement

The ability to move laterally within the cloud environment through poisoned model deployment underscores the complexity and interconnectedness of modern AI and ML infrastructures. Once access is gained, the attacker’s movement is not confined to a single point of entry but can extend across various linked components. This lateral movement increases the scope of potential damage significantly, as attackers can propagate their access to other valuable data and service clusters. Consequently, robust defense mechanisms, such as persistent monitoring of model deployments and strict auditing of permissions, are indispensable in maintaining the integrity and security of AI environments.

Addressing the Vulnerabilities and Mitigations

Google’s Response and Remediation

In response to Unit 42’s disclosure of these vulnerabilities, Google acted promptly to address the security issues. The swift remediation of the flaws emphasizes the importance of quick reaction to discovered vulnerabilities in complex AI systems. Google’s efforts to patch these vulnerabilities indicate their commitment to maintaining the security and reliability of their AI platforms. However, this incident serves as a reminder of the constant vigilance required in managing sophisticated tech infrastructures.

Google’s commitment to strengthening the security of Vertex AI involves implementing preventive measures to mitigate the risks of similar vulnerabilities in the future. These measures include enhanced controls over model deployments and tighter auditing of permission requirements. By refining these security protocols, Google aims to minimize the possibility of unauthorized access and data exfiltration, ensuring that the platform remains a secure environment for its users.

Recommendations for Organizations

To prevent exploitation and potential data theft, organizations using AI and ML platforms must prioritize stringent security measures. Implementing robust defense mechanisms is crucial to safeguarding sensitive data and preventing unauthorized access. Restricting model deployment permissions is one such measure that can significantly reduce security risks. Additionally, closely monitoring and auditing ML workflows helps detect and respond to suspicious activities promptly.

Moreover, organizations should conduct regular security assessments and vulnerability scans on their AI and ML environments. This proactive approach helps identify potential weaknesses and rectify them before they can be exploited by malicious actors. Fostering a security-conscious culture within the organization, coupled with continuous education on best practices, further enhances the overall defense against emerging threats in AI infrastructures.

OpenAI’s ChatGPT Sandbox and Security Considerations

Mozilla’s Findings on ChatGPT Sandbox

Separately, Mozilla’s 0Day Investigative Network (0Din) uncovered intriguing interactions within OpenAI’s ChatGPT sandbox. Users have the ability to upload and execute Python scripts, move files, and download the LLM’s playbook within the sandbox environment. However, OpenAI has classified these interactions as valid functionalities rather than security vulnerabilities. This distinction hinges on the fact that all activities occur within the sandbox’s controlled parameters, designed to provide a secure environment for such actions.

While these functionalities are intended and contained within the sandbox, they prompt discussions about the balance between functionality and security in AI systems. Ensuring that advanced interactive capabilities, like those observed in the ChatGPT sandbox, do not become an avenue for exploitation requires careful design and rigorous testing. The insights from Mozilla’s findings highlight the necessity of maintaining strict control over the boundaries of such functionalities to prevent unexpected security implications.

Balancing Functionality and Security

Recently, Palo Alto Networks’ Unit 42 discovered two major security vulnerabilities in Google’s Vertex AI machine learning (ML) platform. These vulnerabilities pose significant risks, as they could potentially lead to unauthorized access, privilege escalation, and critical data exfiltration if exploited. Vertex AI, which was launched in May 2021, provides a scalable environment for training and deploying custom ML models and AI applications. This discovery is particularly significant due to the growing reliance on platforms like Vertex AI for advancing AI solutions across various sectors. Given the platform’s widespread use, ensuring the security and integrity of such ML services is essential to prevent potential breaches that could have far-reaching consequences on data privacy and system integrity. The discovery by Unit 42 serves as a crucial reminder of the importance of continuously monitoring and securing machine learning platforms to protect them against emerging threats in an increasingly digital and interconnected world.

Explore more