News

CyberArk unveils breakthrough open-source tool that helps organizations safeguard against AI model jailbreaks

News | 24.12.2024

CyberArk, a leader in identity security, has launched FuzzyAI, an open-source framework capable of jailbreaking major AI models. FuzzyAI helps detect vulnerabilities like guardrail bypassing and harmful outputs in AI systems.

AI models are transforming industries with innovative applications in customer interactions, internal process improvements and automation. Internal usage of these models also presents new security challenges for which most organizations are unprepared. >>More

 FuzzyAI helps solve some of these challenges by offering organizations a systematic approach to testing AI models against various adversarial inputs, uncovering potential weak points in their security systems and making AI development and deployment safer. At the heart of FuzzyAI is a powerful fuzzer – a tool that reveals software defects and vulnerabilities – capable of exposing vulnerabilities found via more than ten distinct attack techniques, from bypassing ethical filters to exposing hidden system prompts.

Key features of FuzzyAI include:

  • Comprehensive Fuzzing: FuzzyAI probes AI models with various attack techniques to expose vulnerabilities like bypassing guardrails, information leakage, prompt injection or harmful output generation.
  • An Extensible Framework: Organizations and researchers can add their own attack methods to tailor tests for domain-specific vulnerabilities.
  • Community Collaboration: A growing community-driven ecosystem ensures continuous adversarial techniques and defense mechanisms advancements.

 FuzzyAI empowers organizations and researchers to identify weaknesses and actively fortify their AI systems against emerging threats.