Ptechhub
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs
No Result
View All Result
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs
No Result
View All Result
PtechHub
No Result
View All Result

Popular LLMs dangerously vulnerable to iterative attacks, says Cisco | Computer Weekly

By Computer Weekly by By Computer Weekly
November 7, 2025
Home Uncategorized
Share on FacebookShare on Twitter


Some of the world’s most widely used open-weight generative AI (GenAI) services are profoundly susceptible to so-called “multi-turn” prompt injection or jailbreaking cyber attacks, in which a malicious actor is able to coax large language models (LLMs) into generating unintended and undesirable responses, according to a research paper published by a team at networking giant Cisco.

Cisco’s researchers tested Alibaba Qwen3-32B, Mistral Large-2, Meta Llama 3.3-70B-Instruct, DeepSeek v3.1, Zhipu AI GLM-4.5-Air, Google Gemma-3-1B-1T, Microsoft Phi-4, and OpenAI GPT-OSS-2-B, engineering multiple scenarios in which the various models’ output disallowed content, with success rates ranging from 25.86% against Google’s model, up to 92.78% in the case of Mistral.

The report’s authors, Amy Chang and Nicholas Conley, alongside contributors Harish Santhanalakshmi Ganesan and Adam Swanda, said this represented a two to tenfold increase over single-turn baselines.

“These results underscore a systemic inability of current open-weight models to maintain safety guardrails across extended interactions,” they said.

“We assess that alignment strategies and lab priorities significantly influence resilience: capability-focused models such as Llama 3.3 and Qwen 3 demonstrate higher multi-turn susceptibility, whereas safety-oriented designs such as Google Gemma 3 exhibit more balanced performance.

“The analysis concludes that open-weight models, while crucial for innovation, pose tangible operational and ethical risks when deployed without layered security controls … Addressing multi-turn vulnerabilities is essential to ensure the safe, reliable and responsible deployment of open-weight LLMs in enterprise and public domains.”

What is a multi-turn attack?

Multi-turn attacks take the form of iterative “probing” of an LLM to expose systemic weaknesses that are usually masked because models can better detect and reject isolated adversarial requests.

Such an attack could begin with an attacker making benign queries to establish trust, before subtly introducing more adversarial requests to accomplish their actual goals.

Prompts may be framed with terminology such as “for research purposes” or “in a fictional scenario”, and attackers may ask the models to engage in roleplay or persona adoption, introduce contextual ambiguity or misdirection, or to break down information and reassemble it – among other tactics.

Whose responsibility?

The researchers said their work underscored the susceptibility of LLMs to adversarial attacks and that this was a source of particular concern given all of the models tested were open-weight, which in layman’s terms means anybody who cares to do so is able to download, run and even make changes to the model.

They highlighted as an area of particular concern three more susceptible models – Mistral, Llama and Qwen – which they said had probably been shipped with the expectation that developers would add guardrails themselves, compared with Google’s model, which was most resistant to multi-turn manipulation, or OpenAI’s and Zhipu’s, which both rejected multi-turn attempts more than 50% of the time.

“The AI developer and security community must continue to actively manage these threats – as well as additional safety and security concerns – through independent testing and guardrail development throughout the lifecycle of model development and deployment in organisations,” they wrote.

“Without AI security solutions – such as multi-turn testing, threat-specific mitigation and continuous monitoring – these models pose significant risks in production, potentially leading to data breaches or malicious manipulations,” they added.



Source link

By Computer Weekly

By Computer Weekly

Next Post
Enghouse Acquires Sixbell’s Telco Division

Enghouse Acquires Sixbell's Telco Division

Recommended.

Yeastar premia i partner di tutto il mondo per l’eccellenza e l’innovazione agli Yeahs Awards 2025

Yeastar premia i partner di tutto il mondo per l’eccellenza e l’innovazione agli Yeahs Awards 2025

November 3, 2025
VMware’s New Cloud Program Is For Partners That ‘Fight’ In Broadcom’s ‘Biggest Bet’ Yet, VP Says

VMware’s New Cloud Program Is For Partners That ‘Fight’ In Broadcom’s ‘Biggest Bet’ Yet, VP Says

October 2, 2025

Trending.

Google Sues 25 Chinese Entities Over BADBOX 2.0 Botnet Affecting 10M Android Devices

Google Sues 25 Chinese Entities Over BADBOX 2.0 Botnet Affecting 10M Android Devices

July 18, 2025
Stocks making the biggest moves premarket: Salesforce, American Eagle, Hewlett Packard Enterprise and more

Stocks making the biggest moves premarket: Salesforce, American Eagle, Hewlett Packard Enterprise and more

September 4, 2025
Wesco Declares Quarterly Dividend on Common Stock

Wesco Declares Quarterly Dividend on Common Stock

December 1, 2025
⚡ THN Weekly Recap: New Attacks, Old Tricks, Bigger Impact

⚡ THN Weekly Recap: New Attacks, Old Tricks, Bigger Impact

March 10, 2025
Bloody Wolf Targets Uzbekistan, Russia Using NetSupport RAT in Spear-Phishing Campaign

Bloody Wolf Targets Uzbekistan, Russia Using NetSupport RAT in Spear-Phishing Campaign

February 9, 2026

PTechHub

A tech news platform delivering fresh perspectives, critical insights, and in-depth reporting — beyond the buzz. We cover innovation, policy, and digital culture with clarity, independence, and a sharp editorial edge.

Follow Us

Industries

  • AI & ML
  • Cybersecurity
  • Enterprise IT
  • Finance
  • Telco

Navigation

  • About
  • Advertise
  • Privacy & Policy
  • Contact

Subscribe to Our Newsletter

  • About
  • Advertise
  • Privacy & Policy
  • Contact

Copyright © 2025 | Powered By Porpholio

No Result
View All Result
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs

Copyright © 2025 | Powered By Porpholio