Open-Source AI Models Exposed to Critical Security Vulnerabilities

Researchers recently uncovered over three dozen security vulnerabilities within various open-source artificial intelligence (AI) and machine learning (ML) models. This discovery highlights significant security concerns, some of which could result in remote code execution and information theft. The flaws were identified in widely-used tools such as ChuanhuChatGPT, Lunary, and LocalAI, and were reported through Protect AI’s Huntr bug bounty platform. This raises serious questions about the security measures currently in place and underscores the need for ongoing vigilance.

Major Security Flaws in Popular AI/ML Tools

Lunary’s Critical Vulnerabilities

The most severe vulnerabilities were found within Lunary, a production toolkit for large language models (LLMs). One such vulnerability is the Insecure Direct Object Reference (IDOR) flaw, identified as CVE-2024-7474, which has been rated with a CVSS score of 9.1. This serious vulnerability allows authenticated users to view or delete external users, posing risks of unauthorized access and potential data loss. The improper access control flaw, CVE-2024-7475, also rated at 9.1, further exacerbates the situation by permitting attackers to update SAML configurations. This capability enables unauthorized access to sensitive information and can have far-reaching consequences.

Another significant vulnerability within Lunary, identified as CVE-2024-7473, is yet another IDOR flaw. This particular vulnerability allows unauthorized updates to user prompts by manipulating user-controlled parameters. Such weaknesses highlight the critical need for robust validation and access control mechanisms to prevent exploitation. The presence of multiple IDOR flaws within Lunary underscores a pattern of vulnerabilities that could be targeted by malicious actors, increasing the urgency for implementing comprehensive security measures.

Vulnerabilities in ChuanhuChatGPT

In ChuanhuChatGPT, a path traversal vulnerability, identified as CVE-2024-5982 and carrying a CVSS score of 9.1, has raised concerns. This vulnerability affects the user upload feature and could lead to arbitrary code execution, directory creation, and exposure of sensitive data. Path traversal issues can permit attackers to manipulate file paths and access directories and files outside of the web root folder, further amplifying the risk landscape. The consequences of such vulnerabilities are grave as they jeopardize the confidentiality, integrity, and availability of the system.

The discovery of this path traversal flaw illustrates the need for stringent security measures during the development and deployment phases. Effective measures could have been implemented to ensure proper validation and sanitization of user inputs, thereby mitigating the risk of such vulnerabilities. The implications extend beyond immediate exploitation, as successful attacks can provide a foothold for further malicious activities, demonstrating the cascading effects of security oversights.

Security Issues in LocalAI and Other Tools

LocalAI’s Configuration Handling Flaws

LocalAI, an open-source project for self-hosted LLMs, is susceptible to significant security issues, including one in its configuration file handling, identified as CVE-2024-6983, carrying a CVSS score of 8.8. This vulnerability could enable arbitrary code execution, posing a substantial risk to users relying on this tool for various applications. Configuration files are crucial for the proper functioning of software, and any vulnerabilities in their handling can have widespread implications. The arbitrary nature of potential code execution means attackers could achieve unauthorized control, leading to a range of malicious activities.

Additionally, another issue in LocalAI, identified as CVE-2024-7010 with a CVSS score of 7.5, facilitates API key guessing through timing attacks. Timing attacks exploit the time delay in system responses to deduce sensitive information, such as API keys. The presence of such vulnerabilities necessitates the implementation of more sophisticated security measures, including robust authentication mechanisms and proper rate-limiting on API endpoints to thwart potential attackers.

Other Notable Vulnerabilities

A remote code execution vulnerability in the Deep Java Library (DJL) related to an arbitrary file overwrite bug, rooted in the package’s untar function, has been disclosed and is identified as CVE-2024-8396 with a CVSS score of 7.8. Such a flaw underscores the importance of securing even the most basic functions to prevent severe consequences. Similarly, NVIDIA has released patches for a path traversal flaw in its NeMo generative AI framework, identified as CVE-2024-0129 with a CVSS score of 6.3, which could lead to code execution and data tampering. Addressing these vulnerabilities is crucial to maintaining the integrity and security of AI/ML frameworks.

These discoveries highlight the persistent and evolving nature of security challenges within open-source AI and ML models. To counter these emerging threats, users are advised to update their AI/ML installations to the latest versions to secure their systems and mitigate potential attacks. Proactive measures, such as regular security assessments and the implementation of advanced defensive tools, are imperative to stay ahead of malicious exploitation attempts.

Proactive Measures and New Security Tools

Introduction of Vulnhuntr by Protect AI

In response to these emerging security threats, Protect AI has announced the release of Vulnhuntr, an open-source Python static code analyzer designed to identify zero-day vulnerabilities in Python codebases. Vulnhuntr leverages large language models (LLMs) to pinpoint potential vulnerabilities by examining files most likely to handle user input. The tool then traces the function call chain from input to output for a comprehensive final analysis. This proactive approach aims to enhance the security posture of organizations by offering a means to identify and remediate vulnerabilities before they can be exploited.

Vulnhuntr represents a significant advancement in the field of static code analysis, providing developers and security professionals with a powerful tool tailored to the unique challenges posed by AI and ML models. The ability to detect zero-day vulnerabilities is particularly valuable as it addresses weaknesses that have not yet been disclosed or patched. By incorporating machine learning techniques, Vulnhuntr enhances the accuracy and efficiency of vulnerability detection, thereby reducing the window of opportunity for attackers.

New Jailbreak Techniques by Mozilla’s 0Din

Moreover, a new jailbreak technique by Mozilla’s 0Day Investigative Network (0Din) reveals that malicious prompts encoded in hexadecimal and emojis can bypass OpenAI ChatGPT’s safeguards. This technique exploits linguistic loopholes by instructing the model to perform hex conversion without recognizing potentially harmful outcomes. The discovery of this method underscores the need for continuous improvement and adaptation of security mechanisms to address novel exploitation techniques.

The ability to bypass AI safeguards using such innovative methods poses a unique challenge to developers and security researchers. It highlights the complex interplay between natural language processing capabilities and security measures, necessitating a multi-faceted approach to threat mitigation. The findings by 0Din emphasize the importance of rigorous testing and validation of AI models to uncover and address potential vulnerabilities in their operational logic.

Conclusion

Researchers have recently discovered over 36 security vulnerabilities in various open-source artificial intelligence (AI) and machine learning (ML) models, raising significant security concerns. These vulnerabilities have the potential to allow for remote code execution and theft of sensitive information. They were found in widely-used tools like ChuanhuChatGPT, Lunary, and LocalAI, among others. The issues were reported through Protect AI’s Huntr bug bounty platform, a system designed to encourage the discovery and reporting of security flaws. This development raises serious questions about the adequacy of current security measures in these tools and highlights the pressing need for continuous vigilance and improvement in AI and ML security. The findings underscore the importance of proactive measures to secure these models, as they play an increasingly crucial role in various applications and industries. The identification of these flaws is a reminder that, despite the rapid advancements in AI and ML, the security of these technologies must not be overlooked. Ensuring robust security protocols and regular audits can help mitigate potential risks associated with these vulnerabilities.

Explore more

Why Are Companies Suddenly Hiring Again in 2026?

The sudden ping of a LinkedIn notification or a direct recruiter email has recently transformed from a rare digital relic into a daily occurrence for many professionals. After a prolonged period characterized by “ghost” job postings and a deafening silence from human resources departments, the professional landscape has reached a startling tipping point. In a single month, U.S. job openings

HR Leadership Is Crucial for Successful AI Transformation

The rapid integration of artificial intelligence into the modern corporate landscape is no longer a futuristic prediction but a present-day reality, fundamentally reshaping how organizations operate, hire, and plan for the future. In today’s market, 95% of C-suite executives identify AI as the most significant catalyst for transformation they will witness in their entire professional lives. This shift represents a

Does Your Response Speed Signal Your Professional Status?

When an incoming notification pings on a high-resolution smartphone screen, the decision to let it sit for hours rather than seconds is rarely a matter of simple forgetfulness. In the contemporary corporate landscape, an employee who responds to every message within the blink of an eye is often lauded as a dedicated team player, yet in many elite professional circles,

How AI-Native Architecture Will Power 6G Wireless Networks

The fundamental transformation of global telecommunications is no longer defined by incremental increases in bandwidth but by the total integration of cognitive computing into the very fabric of signal transmission. As of 2026, the industry is witnessing the sunset of the era where Artificial Intelligence functioned merely as an external troubleshooting tool for cellular towers. Instead, the groundwork for 6G

The Global Race Toward 6G Engineering and Commercial Reality

The relentless momentum of global telecommunications has reached a pivotal juncture where the transition from laboratory theory to tangible engineering hardware defines the current technological landscape. If every decade of telecommunications has a “north star,” the year 2030 is currently pulling the entire global engineering community toward its orbit with an irresistible force. We are currently navigating a critical three-year