A new jailbreak vulnerability in OpenAI’s ChatGPT-4o, dubbed “Time Bandit,” has been exploited to bypass the chatbot’s built-in safety functions.
This vulnerability allows attackers to manipulate the chatbot into producing illicit or dangerous content, including instructions for malware creation, phishing scams, and other malicious activities.
The exploitation of this jailbreak has raised alarms within the cybersecurity community due to its potential to be scaled up for malign purposes by threat actors.
The “Time Bandit” jailbreak uncovered by researcher Dave Kuszmar confuses the AI by anchoring its responses to a specific historical period. Attackers can exploit this vulnerability in two primary ways: through direct interaction with the AI or by utilizing the Search functionality integrated into ChatGPT.
Collect Threat Intelligence with TI Lookup to Improve Your Company’s Security - Get 50 Free Request
The bug was first reported by cybersecurity researcher Dave Kuszmar and documented by the CERT Coordination Center (CERT/CC). During controlled testing, they were able to replicate the jailbreak multiple times. Once initiated, ChatGPT sometimes produced illicit content even after detecting and removing specific prompts that violated usage policies.
Most notably, the jailbreak proved more effective when historical time frames from the 1800s and 1900s were used.
While exploiting the vulnerability through direct prompts did not require user authentication, utilizing the Search feature did require a logged-in account. This dual method of exploitation demonstrates the versatility of the “Time Bandit” vulnerability.
The implications of this vulnerability are far-reaching. By bypassing OpenAI’s strict safety guidelines, attackers could leverage ChatGPT to generate step-by-step instructions for creating weapons, drugs, or malware.
It could also be used to mass-produce phishing scams, social engineering scripts, or other harmful content.
Using a legitimate and widely trusted tool like ChatGPT further obfuscates malicious activities, making detection and prevention more challenging.
Experts warn that under the control of organized cybercriminals, “Time Bandit” could facilitate large-scale malicious operations, posing a significant threat to cybersecurity and public safety.
OpenAI has already acted to address the vulnerability. In a statement, an OpenAI spokesperson emphasized the company’s commitment to safety: “It is very important to us that we develop our models safely. We don’t want our models to be used for malicious purposes. We appreciate you for disclosing your findings. We’re constantly working to make our models safer and more robust against exploits, including jailbreaks, while also maintaining the models’ usefulness and task performance.”
The recent trading DeepSeek R1 model was also Jailbroken to Generate Ransomware Development Scripts, “DeepSeek R1 provided detailed instructions and generated a malicious script designed to extract credit card data from specific browsers and transmit it to a remote server.”
For Daily Security Updates! Follow us on Google News, LinkedIn, and X