
Researchers recently uncovered over three dozen security vulnerabilities within various open-source artificial intelligence (AI) and machine learning (ML) models. This discovery highlights significant security concerns, some of which could result in remote code execution and information theft. The flaws were identified in widely-used tools such as ChuanhuChatGPT, Lunary, and LocalAI, and were reported through Protect AI’s Huntr bug bounty platform.