Sunday , August 31 2025
Qwen

ChatGPT, DeepSeek, Qwen 2.5-VL Vulnerable to AI Jailbreaks

This week, multiple research teams showcased jailbreaks for popular AI models, including OpenAI’s ChatGPT, DeepSeek, and Alibaba’s Qwen.

After its launch, the open-source R1 model by Chinese company DeepSeek caught the attention of the cybersecurity industry. Experts found that jailbreak methods, previously patched in other AI models, still function against DeepSeek.

“SikkahBot” Malware targets “bKash” “Nagad” “MYGP” “DBBL” with banking users in Bangladesh

A new Android malware called SikkahBot is targeting students in Bangladesh by pretending to be official apps from the Bangladesh...
Read More
“SikkahBot” Malware targets “bKash” “Nagad” “MYGP” “DBBL” with banking users in Bangladesh

F5 Executive Forum in Dhaka Explores App Delivery, Security, and AI Challenges

As organizations embrace digital transformation, the complexity of managing applications across Hybrid, Multicloud, and AI-driven environments continues to grow. Than...
Read More
F5 Executive Forum in Dhaka Explores App Delivery, Security, and AI Challenges

PromptLock: The First AI-Driven Ransomware Appears

AI-driven malware has emerged sooner than anticipated, with the first identified ransomware using AI for local tasks now found. ESET...
Read More
PromptLock: The First AI-Driven Ransomware Appears

CVE-2025-55177
WhatsApp patches zero day vuln exploited in the wild

Meta's WhatsApp Security Team has fixed a zero-day vulnerability (CVE-2025-55177) in WhatsApp for iOS (before v2.25.21.73), WhatsApp Business for iOS...
Read More
CVE-2025-55177  WhatsApp patches zero day vuln exploited in the wild

3 critical vulnerabilities affect Hikvision product: Patch Now

The Hikvision Security Response Center issued advisory revealing three critical vulnerabilities in HikCentral products. CVE identifiers CVE-2025-39245, CVE-2025-39246, and CVE-2025-39247...
Read More
3 critical vulnerabilities affect Hikvision product: Patch Now

Salt Typhoon To Exploit Cisco, Palo Alto, Ivanti Flaws to Breach 600 Org Globally

The China-linked APT group Salt Typhoon is still attacking networks globally, focusing on telecommunications, government, transportation, hospitality, and military sectors....
Read More
Salt Typhoon To Exploit Cisco, Palo Alto, Ivanti Flaws to Breach 600 Org Globally

Storm-0501 Deletes Data and Backups Post-Exfiltration on Azure in Hybrid Cloud Attacks

Storm-0501 has erased data and backups after stealing information from a victim's Microsoft Azure environment in a new cloud based...
Read More
Storm-0501 Deletes Data and Backups Post-Exfiltration on Azure in Hybrid Cloud Attacks

Breaking the Passkey: SquareX Discloses Major Passkey Vulnerability at DEF CON 33

It is no secret that passwords are highly susceptible to phishing and brute force attacks. This led to the mass...
Read More
Breaking the Passkey: SquareX Discloses Major Passkey Vulnerability at DEF CON 33

(CVE-2025-20241)
Cisco Warns of High-Severity Flaw in Nexus Switches

Cisco Systems released a security advisory about a critical denial-of-service vulnerability in the Nexus 3000 and 9000 Series Switches using...
Read More
(CVE-2025-20241)  Cisco Warns of High-Severity Flaw in Nexus Switches

Malaysia Launches World’s First AI-powered Bank

Malaysia is boosting its digital transformation by launching its first fully AI-powered banking service. This initiative, which comes just before...
Read More
Malaysia Launches World’s First AI-powered Bank

AI jailbreaking allows attackers to bypass safeguards designed to stop LLMs from producing harmful content. Security researchers have demonstrated that methods like prompt injection and model manipulation can overcome these protections.

Threat intelligence firm Kela found that DeepSeek is affected by Evil Jailbreak, where a chatbot is made to act as an evil confidant, and Leo, which allows the chatbot to take on an unrestricted persona. ChatGPT has fixed these vulnerabilities.

Palo Alto Networks’ Unit42 reported that DeepSeek is vulnerable to known AI jailbreak techniques.

The security firm successfully conducted the attack known as Deceptive Delight, which tricks generative AI models by embedding unsafe or restricted topics in benign narratives. This method was tested in the fall of 2024 against eight LLMs with an average success rate of 65%.

Palo Alto has successfully executed the Bad Likert Judge jailbreak, which asks the LLM to evaluate the harmfulness of responses using a Likert scale and generate examples that fit the scale.

Researchers discovered that DeepSeek is vulnerable to Crescendo, a jailbreak method that begins with innocuous dialogue and gradually shifts towards forbidden topics.

Alibaba has announced a new version of its Qwen AI model, claiming it outperforms the DeepSeek model.

Kela announced on Thursday that Alibaba’s new Qwen 2.5-VL model has vulnerabilities similar to those recently discovered in DeepSeek.

Researchers at a threat intelligence firm found that jailbreaks designed for DeepSeek also work on Qwen. They successfully tested an existing jailbreak called Grandma, which tricks the model into sharing dangerous information by having it role-play as a grandmother.

Kela found that Qwen 2.5-VL created content about developing ransomware and other malware.

“The ability of AI models to produce infostealer malware instructions raises serious concerns, as cybercriminals could leverage these capabilities to automate and enhance their attack methodologies,” Kela said.

Many jailbreak methods for ChatGPT have been fixed over the years, but researchers still discover new ways to bypass its protections.

CERT/CC reported that researcher Dave Kuszmar found a ChatGPT-4o jailbreak vulnerability called Time Bandit. This vulnerability allows users to ask the AI about specific historical events or instruct it to pretend to assist in such events.

“The jailbreak can be established in two ways, either through the Search function, or by prompting the AI directly,” CERT/CC explained in an advisory. “Once this historical timeframe has been established in the ChatGPT conversation, the attacker can exploit timeline confusion and procedural ambiguity in following prompts to circumvent the safety guidelines, resulting in ChatGPT generating illicit content. This information could be leveraged at scale by a motivated threat actor for malicious purposes.”

Check Also

Passkey

Breaking the Passkey: SquareX Discloses Major Passkey Vulnerability at DEF CON 33

It is no secret that passwords are highly susceptible to phishing and brute force attacks. …