Researchers Warn of Privilege Escalation Risks in Google’s Vertex AI ML Platform

Researchers Warn of Privilege Escalation Risks in Google's Vertex AI ML Platform

November 15, 2024 at 08:30AM

Cybersecurity researchers uncovered two vulnerabilities in Google’s Vertex AI platform that could allow exploitation for privilege escalation and data exfiltration. Attackers could manipulate job permissions to access restricted resources and deploy malicious models to extract sensitive information. Google has addressed these issues, urging organizations to implement stricter model deployment controls.

### Meeting Takeaways – Nov 15, 2024

1. **Security Vulnerabilities in Google Vertex AI:**
– Cybersecurity experts identified two significant security flaws in Google’s Vertex AI machine learning platform.
– These vulnerabilities allow for privilege escalation and potential exfiltration of proprietary models.

2. **Details of Vulnerabilities:**
– **Privilege Escalation via Custom Job Permissions:**
– Attackers can exploit these permissions to access all project data services.
– A manipulated custom job pipeline can launch a reverse shell, granting backdoor access to the environment.
– Extensive service account permissions allow access to internal Google Cloud resources, including storage and BigQuery.

– **Deployment of Poisoned Models:**
– A poisoned model can create a reverse shell upon deployment, enabling attackers to enumerate Kubernetes clusters and obtain necessary credentials for executing commands.
– This allows lateral movement from Google Cloud Platform (GCP) to Kubernetes environments.

3. **Potential Consequences:**
– Malicious deployment can lead to full model exfiltration, impacting sensitive data and proprietary information.
– The risk escalates when developers unknowingly deploy compromised models from public repositories.

4. **Recommendations for Organizations:**
– Implement strict controls over model deployments.
– Audit permissions required for deploying models in tenant projects.

5. **Related Findings on OpenAI ChatGPT Sandbox:**
– Mozilla’s 0Day Investigative Network reported that it’s possible to interact within OpenAI’s ChatGPT sandbox environment, allowing for the execution of scripts and manipulation of files.
– OpenAI acknowledges these actions as expected behaviors within the sandbox, not security vulnerabilities.

6. **Action Steps Moving Forward:**
– Organizations should prioritize reviewing their AI deployment strategies and security protocols.
– Continuous monitoring and updates are essential to safeguard sensitive machine learning environments.

7. **Engagement Opportunity:**
– Follow for more insights and updates on cybersecurity and AI developments on Twitter and LinkedIn.

Full Article