Researchers Identify Vulnerabilities in Widely Used Open-Source Machine Learning Frameworks

Security Vulnerabilities Discovered in Popular Machine Learning Tools

Recent findings by cybersecurity researchers have unveiled multiple critical vulnerabilities within open-source machine learning frameworks such as MLflow, H2O, PyTorch, and MLeap. These weaknesses could potentially allow unauthorized code execution, posing significant security risks to organizations leveraging these tools.

The vulnerabilities, initially reported by JFrog, are part of a larger suite of security flaws totaling 22 that the supply chain security firm disclosed last month. The newly identified issues focus specifically on ML clients, differing from the previously outlined server-side vulnerabilities. They exist in libraries responsible for processing safe model formats, including Safetensors, which are increasingly used for machine learning tasks.

Attackers exploiting these vulnerabilities can compromise an ML client within an organization, enabling extensive lateral movement across the network. Given that ML clients often interact with critical ML services—such as ML Model Registries and MLOps Pipelines—this creates a pathway for exposing sensitive information, including model registry credentials. Consequently, a malicious actor could backdoor stored ML models or achieve remote code execution, thereby undermining the integrity of the entire ML ecosystem.

The vulnerabilities identified include a critical issue within MLflow that results from insufficient sanitization, allowing a potential cross-site scripting (XSS) attack when running untrusted recipes in Jupyter Notebooks. Another serious concern relates to H2O, where unsafe deserialization when importing untrusted ML models could lead to remote code execution. PyTorch is also affected; its TorchScript feature has a significant path traversal flaw, which might facilitate denial-of-service attacks or arbitrary file overwrites targeting critical system files.

Moreover, MLeap exhibits a path traversal vulnerability that could lead to a Zip Slip attack when loading zipped models, raising further concerns regarding arbitrary file overwrites and potential code execution. Such vulnerabilities necessitate immediate attention from organizations to mitigate risks.

JFrog emphasizes the importance of exercising caution when handling ML models, including those loaded from repositories deemed ‘safe.’ Notably, even models classified as such can introduce substantial risks, including arbitrary code execution. As articulated by Shachar Menashe, JFrog’s VP of Security Research, although AI and machine learning technologies hold great promise, they can also serve as conduits for cyber threats if not managed prudently.

In assessing the risks associated with these vulnerabilities, business owners should be mindful of tactics identified in the MITRE ATT&CK framework, including techniques for initial access and privilege escalation. Addressing these vulnerabilities is critical not only to safeguard intellectual property but also to protect broader organizational assets from potential breaches.

As the landscape of machine learning continues to evolve, maintaining robust security practices will be essential for organizations to navigate these challenges effectively. Awareness and proactive measures will remain key components in defending against the evolving threats posed by malicious actors in the cybersecurity realm.

Source link