LLMs remain vulnerable to malicious jailbreaking

Large language models, LLMs, are impressive. They can teach, draft, translate, code. But they’re still surprisingly easy to manipulate. Research from Michael Fire, Yitzhak Elbazis, Adi Wasenstein, and Lior Rokach at Ben Gurion University shows this clearly. They demonstrated that even commercial AI models with built-in safety systems can be tricked into generating harmful or illegal output. That includes things like bomb-making instructions, insider trading tactics, or how to run a money laundering operation.

The method that bypasses these built-in constraints is known as a “universal jailbreak attack.” Basically, the model is fed a very specific prompt that nudges it out of its careful alignment. The result: it responds like there are no safeguards at all. This isn’t theoretical. It works across multiple, well-known AI systems in real-world conditions. Their research paper, “Dark LLMs: The Growing Threat of Unaligned AI Models,” details how this happens.

If your business depends on AI-generated insights or decision-making, this matters. The assumption that your model is secure, because it has guardrails, is no longer valid. These guardrails are patchable, yes, but jailbreaking exploits something deeper. It leverages the model’s core function: responding to language patterns. With the right pattern, those restrictions break.

For C-suite leaders, the takeaway is simple: Don’t assume your AI tools are safe by default. Audit them. Stress-test them. Decide if the risk they introduce is one you understand and can manage. We’re beyond the point where “ethical AI” is a buzzword. It’s now a business continuity and reputational exposure issue.

Open-source LLMs present a unique, uncontrollable risk

Open-source LLMs offer speed and flexibility. Developers love them. But they’re also nearly impossible to contain once released. According to the same Ben Gurion researchers, these models often come uncensored, with limited or no safety constraints. After release, they’re copied, archived, and shared across servers and devices globally, immediately outside anyone’s control.

The bigger problem is how attackers exploit them. One model can be used to jailbreak another. So even if you think you’ve secured your commercial AI system, it can get queried through structures built on compromised open-source tools. Once that happens, your internal safeguards may be bypassed without you knowing.

Unlike commercial models, which vendors can update or patch, open-source versions don’t have centralized control. Once in the wild, they’re permanent. This creates new vulnerabilities for businesses that integrate open-source AI tools or interface with external models. From a security standpoint, this is an open door. From a compliance standpoint, it’s liability.

C-suite leaders should reconsider any assumption that open-source software is automatically aligned with innovation and cost-efficiency. When it comes to security, lack of control is risk. Understand what your team is deploying. If it’s open-source, treat it with the same scrutiny as any unverified vendor system, because that’s exactly what it becomes once adopted at scale.

A multi-layered defensive strategy is essential

Securing AI systems won’t come from one technical fix. It requires layered defense, across design, deployment, and operations. The research from Ben Gurion University is clear: model safeguards on their own don’t handle jailbreaks effectively. So the answer isn’t just better guardrails. It’s system-wide architecture that anticipates exploitation.

Start with training data. If a model is exposed during training to content related to bomb-making, money laundering, or deepfake manipulation, it retains patterns that can be triggered later. Curation of training data needs to be deliberate, specifically engineered to exclude these risks up front.

Next level is middleware. Tools like IBM’s Granite Guardian and Meta’s Llama Guard prove this is doable. These sit between users and the model, reviewing both prompts and responses in real time. It’s a firewall, not for computers, but for language. When deployed right, this kind of interception reduces exposure dramatically.

Another approach is machine unlearning. Unlike retraining from scratch, this lets a model “forget” targeted information. You fix issues without losing your entire training investment. Then there’s red teaming. Invite adversarial testing. Pay bounties. Publish results. Open your system to constant scrutiny, not just periodic review.

Leaders serious about deploying LLMs at scale should treat safety as a product feature, not a checkbox. System architecture, not just model alignment, determines resilience. And none of this is speculative. There are already known exploits and available tooling to stop them. Invest where the threats are real.

The fundamental nature of LLMs challenges full security

LLMs don’t operate like traditional software. They don’t follow fixed rules. This makes them useful, but also unpredictable. As Justin St-Maurice from Info-Tech Research Group puts it, LLMs are probabilistic. They don’t know what they’re doing, they just calculate what to say next based on patterns.

That means jailbreaks aren’t code hacks. They’re context shifts. A prompt tweaks how the model interprets a request, and suddenly the ethical safeguards fall away. There’s no isolated system to patch, just an open-ended reasoning engine designed to produce plausible text.

The issue isn’t just technical. It’s conceptual. If your company expects 100% safe output from an LLM, you’re gambling against the system’s design. Everything about these models, scale, scope, flexibility, comes from their lack of deterministic constraints. The moment you prioritize creative adaptability, you trade away absolute control.

For C-level executives, the right move is clarity. Know the limitations. Choose where and how to deploy LLMs based on the level of potential harm if something fails. Use middleware. Monitor usage. Disallow direct access where risky. And most importantly, never assume the tool knows what it’s doing. Because it doesn’t.

Urgent regulatory and technical governance is critical

The capabilities of large language models are advancing quickly. The benefits are real, faster research, more efficient operations, new product opportunities. But so are the risks. According to the team at Ben Gurion University, the same tools used to accelerate progress can also be used to produce detailed instructions for criminal activity or disinformation campaigns.

This isn’t a future concern. The misuse is happening now. And the more powerful these systems become, the harder it will be to contain the outcomes without external enforcement and oversight. Technologists can provide tooling, but that’s not enough. Regulation, standards, and public policy need to move in parallel, and fast.

Their recommendation is clear: treat unaligned LLMs as high-risk assets. Control access. Apply age restrictions. Audit deployments. Ensure there’s liability clarity for misuse, intentional or otherwise. These aren’t overreactions. They’re normalized safety frameworks in almost every other high-impact technology sector.

For governments, it’s time to classify and treat unfiltered models with the same seriousness as restricted content. For enterprise leaders, it means taking an active role. Don’t wait for regulation to mature, lead in setting your guardrails internally. Establish governance committees. Partner with technical experts. Define compliance around LLM risk.

The researchers warn that the window for proactive leadership is closing. Models are rapidly improving, while misuse strategies are scaling globally. Without alignment between builders, regulators, and users, the long-term impact could be damaging, economically, politically, and socially.

In any high-consequence technology, leadership means forward action. The responsible move now is to guide how these tools are controlled, applied, and made safe for broad use. What’s at risk isn’t just technical disruption. It’s institutional trust. And once trust breaks, impact is tough to unwind.

Key highlights

  • LLMs are easily jailbroken despite safeguards: Most AI systems can still be manipulated to produce harmful or illegal content through prompt-based jailbreaks. Leaders should not assume built-in filters provide sufficient protection and must evaluate model vulnerability regularly.
  • Open-source LLMs amplify uncontrolled risk: Once uncensored LLMs are released, they become unpatchable and freely sharable, increasing exposure to misuse. Executives should treat open-source deployments with heightened scrutiny and apply containment policies before adoption.
  • A multi-layered defense is business-critical: Relying solely on model-level safety controls is insufficient. Leaders should implement layered defenses, including curated training data, middleware firewalls, machine unlearning, red teaming, and internal governance frameworks.
  • AI’s design limits full safety enforcement: Because LLMs are probabilistic, not rule-based, they can’t reliably distinguish harmful from acceptable contexts. C-suite leaders should plan for continuous oversight and monitoring rather than expect permanent containment.
  • AI governance needs immediate action: Without swift regulatory and policy guidance, the misuse of LLMs is likely to escalate rapidly. Leaders should establish internal AI accountability structures and prepare to align with forthcoming regulatory requirements.

Alexander Procter

June 12, 2025

7 Min