More ChatGPT Jailbreaks Are Evading Safeguards On Sensitive Topics
SOURCE: FORBES.COM
FEB 01, 2025
Contributor
Alex Vakulov is a cybersecurity expert focused on consumer security.
Feb 1, 2025,09:46am EST
Updated Feb 1, 2025, 09:50am EST
Artificial intelligence (AI) chatbots like OpenAI’s ChatGPT and Google’s Gemini are revolutionizing the way users interact with technology. From answering queries and automating tasks to assisting with software development, AI models have become indispensable tools.
However, their increasing capabilities also present significant cybersecurity risks. One recent example is the Time Bandit jailbreak, a flaw in ChatGPT that allows users to bypass OpenAI’s safety measures and extract information on sensitive topics, such as malware creation and weapons development.
While AI models have safeguards in place to prevent misuse, researchers and cybercriminals continuously explore ways to circumvent these protections. The Time Bandit jailbreak highlights a broader issue: AI chatbots are vulnerable to manipulation, posing risks not only to enterprises but also to individual users. Understanding these risks and implementing protective measures is crucial for safe interaction with AI tools and avoiding data leaks.
PROMOTED
The Time Bandit exploit, discovered by cybersecurity researcher David Kuszmar, takes advantage of two fundamental weaknesses in ChatGPT:
By manipulating these weaknesses, users can trick ChatGPT into thinking it is in a different historical period while still using modern knowledge. This enables the AI to generate responses that would normally be restricted, such as instructions on coding polymorphic malware or creating weapons.
A cybersecurity test demonstrated how Time Bandit could deceive ChatGPT into assuming it was assisting a programmer in 1789 while leveraging modern coding practices. The AI, confused by the timeline shift, provided detailed guidance on crafting polymorphic malware, including self-modifying code and execution techniques that would typically be restricted.
While OpenAI has acknowledged the issue and is working on mitigations, the jailbreak still functions in some scenarios, raising concerns about the security of AI-driven chatbots.
Beyond the Time Bandit jailbreak, AI chatbots present several cybersecurity risks that consumers should be aware of:
Forbes Daily: Join over 1 million Forbes Daily subscribers and get our best stories, exclusive reporting and essential analysis of the day’s news in your inbox every weekday.
Get the latest news on special offers, product updates and content suggestions from Forbes and its affiliates.
Sign Up
By signing up, you agree to our Terms of Service, and you acknowledge our Privacy Statement. Forbes is protected by reCAPTCHA, and the Google Privacy Policy and Terms of Service apply.
AI-generated text can be used to craft highly convincing phishing emails or scam messages. Attackers can leverage chatbots to produce flawless, personalized phishing content that deceives victims into revealing sensitive information.
Users often input confidential information into chatbots, assuming their data is secure. However, AI models retain and process input data, which can be a privacy risk if exposed through security breaches or model training data leaks.
Bad actors can use AI chatbots to spread misinformation or generate harmful content, making it harder for users to discern between real and fake information online.
As demonstrated by the Time Bandit jailbreak, AI can be manipulated to generate harmful code or assist in cybercriminal activities. While safeguards exist, they are not foolproof.
Many chatbots integrate with external services via plugins and APIs. A compromised third-party service can introduce security risks, leading to unauthorized access or data leaks.
Given these risks, you must take proactive steps to ensure your security while interacting with AI chatbots. Here are some best practices:
Avoid sharing sensitive data such as passwords, financial details, or confidential business information with AI chatbots. Assume that any data inputted could be stored or accessed later.
Do not rely on AI-generated responses for critical decision-making without verification. If using AI for research, cross-check the information from credible sources.
If you notice prompts or conversations that appear to bypass AI safeguards, report them to the chatbot provider. The ethical use of AI helps maintain security for all users.
Attackers may use AI chatbots to spread malicious links. Before clicking on links or downloading files suggested by AI, verify their legitimacy using cybersecurity tools.
Stick to AI models from reputable providers with clear privacy policies and regular security updates. Avoid unknown or unverified AI tools that could pose greater risks.
Ensure that your web browser, security software, and any AI-related apps are up to date to mitigate known vulnerabilities.
Follow me on Twitter or LinkedIn.
Follow
Alex Vakulov is a cybersecurity expert with over 20 years of experience in malware analysis and removal. He writes for various tech publications
LATEST NEWS
WHAT'S TRENDING
Data Science
5 Imaginative Data Science Projects That Can Make Your Portfolio Stand Out
OCT 05, 2022
SOURCE: THE-DECODER.COM
MAR 22, 2025
SOURCE: MSN.COM
MAR 15, 2025
SOURCE: UK.NEWS.YAHOO.COM
MAR 08, 2025
SOURCE: TECHRADAR.COM
MAR 01, 2025