GPT-5 Jailbroken in 24 Hours, Exposing AI Security Risks

Independent researchers jailbroke OpenAI's GPT-5 within 24 hours of its release, exposing vulnerabilities in safeguards that allow harmful content and data leaks. This raises serious doubts about its enterprise readiness amid privacy and compliance risks. Experts urge enhanced security measures to ensure safe AI adoption.
GPT-5 Jailbroken in 24 Hours, Exposing AI Security Risks
Written by Sara Donnelly

Just days after OpenAI unveiled its highly anticipated GPT-5 model, independent security researchers have sounded the alarm on its vulnerabilities, raising serious questions about its readiness for corporate deployment. Red teams, groups of ethical hackers tasked with probing AI systems for weaknesses, managed to “jailbreak” the model—bypassing its built-in safeguards—in under 24 hours. This swift breach highlights ongoing challenges in AI security, where advanced language models like GPT-5 are expected to handle sensitive enterprise tasks but falter under targeted attacks.

The jailbreaks exposed flaws in GPT-5’s context handling and guardrail enforcement, allowing researchers to manipulate the model into generating harmful content or revealing restricted information. One firm, Lakera AI, detailed how simple prompts could override safety measures, while another, Adversa AI, replicated the feats with similar ease. These findings echo earlier incidents with predecessors like GPT-4, but the speed of compromise here suggests minimal progress in fortifying defenses.

Rapid Exploits and Enterprise Risks

According to a report from SecurityWeek, the red teams’ assessments paint a grim picture for businesses eyeing GPT-5 integration. Lakera AI’s researchers noted that the model’s vulnerabilities make it “nearly unusable” in high-stakes environments, where data privacy and compliance are paramount. For instance, in enterprise settings, AI tools must adhere to regulations like GDPR or HIPAA, yet GPT-5’s porous barriers could lead to unintended data leaks or biased outputs that invite legal scrutiny.

Adversa AI’s parallel tests corroborated these concerns, demonstrating how adversaries could exploit the model’s reasoning capabilities to extract proprietary code or simulate phishing attacks. This isn’t isolated; historical breaches, such as the 2023 ChatGPT data exposure confirmed by SecurityWeek, underscore a pattern of recurring weaknesses in OpenAI’s ecosystem. Industry insiders worry that without robust fixes, companies might face reputational damage or regulatory fines from deploying such flawed technology.

Broader Implications for AI Adoption

The backlash extends beyond technical glitches to OpenAI’s development priorities. Critics argue that the rush to release GPT-5, amid competitive pressures from rivals like Google’s Gemini or Anthropic’s Claude, may have sidelined thorough security vetting. A recent article in InfotechLead highlighted user outrage over misleading marketing claims, with CEO Sam Altman promising enterprise-grade capabilities that real-world tests have debunked.

Moreover, red teaming efforts, while commendable, appear insufficient. OpenAI has invited external experts for safety testing, as noted in a Metaverse Post piece from earlier this year, yet the quick jailbreaks suggest gaps in this process. For enterprises, this means weighing the model’s advanced coding and workflow features against potential security pitfalls, possibly delaying widespread adoption.

Path Forward: Calls for Reform

In response, OpenAI has pledged updates, including doubled rate limits for paid users to address performance complaints, per a BleepingComputer report. However, experts like those from Concentric AI, in their 2025 guide on ChatGPT risks, emphasize the need for layered defenses, such as continuous monitoring and adversarial training, to make models like GPT-5 viable for business use.

Ultimately, these revelations could reshape how companies approach AI integration, pushing for more transparent evaluations and collaborative red teaming across the industry. As breaches mount—evidenced by recent incidents like Google’s Salesforce hack detailed in SecurityWeek—the onus is on developers to prioritize security, ensuring that innovation doesn’t come at the cost of trust. For now, GPT-5’s enterprise promise remains overshadowed by its vulnerabilities, urging caution among potential adopters.

Subscribe for Updates

EnterpriseSecurity Newsletter

News, updates and trends in enterprise-level IT security.

By signing up for our newsletter you agree to receive content related to ientry.com / webpronews.com and our affiliate partners. For additional information refer to our terms of service.

Notice an error?

Help us improve our content by reporting any issues you find.

Get the WebProNews newsletter delivered to your inbox

Get the free daily newsletter read by decision makers

Subscribe
Advertise with Us

Ready to get started?

Get our media kit

Advertise with Us