AI
AI

Cisco Alerts: Optimization of LLMs Could Create Security Threats

Photo credit: venturebeat.com

The rapid evolution of AI technology is redefining cybersecurity approaches, particularly through the emergence of weaponized large language models (LLMs) fine-tuned for offensive operations. This trend poses significant challenges for Chief Information Security Officers (CISOs), compelling them to rethink their security strategies.

LLMs such as FraudGPT, GhostGPT, and DarkGPT are now available for relatively low subscription fees, starting at around $75 per month. These models are specifically optimized for various cyberattack methodologies, including phishing, exploit development, code obfuscation, vulnerability scanning, and credit card validation.

Criminal organizations, hacktivists, and nation-state actors are increasingly capitalizing on the financial avenues that weaponized LLMs present, providing access to these tools in a manner akin to legitimate software as a service (SaaS) offerings. Subscription packages frequently encompass access to user-friendly dashboards, APIs, ongoing updates, and, in some instances, customer support.

The evolution of weaponized LLMs is being meticulously monitored by sources like VentureBeat, revealing a disturbing trend where the distinction between legitimate development platforms and cybercriminal toolkits is fading. As leasing options for these models become more economical, an increasing number of individuals are experimenting with these dangerous technologies, marking a critical shift in the threat landscape driven by AI.

Legitimate LLMs in the Crosshairs

The rapid proliferation of weaponized LLMs places legitimate models in peril of being exploited within cybercriminal frameworks. The stark reality is that these legitimate tools are now in jeopardy of being hijacked for malicious activities.

Evidently, the efficacy with which an LLM is fine-tuned correlates directly with its propensity to produce harmful outputs. According to Cisco’s State of AI Security Report, fine-tuned LLMs are twenty-two times more likely to generate unsafe content than their baseline counterparts. Although fine-tuning is critical for increasing contextual accuracy, this process also compromises built-in safety measures, thereby increasing susceptibility to exploits like jailbreaks and prompt injections.

Findings from Cisco illustrate that as LLMs become more suitable for production, they concurrently become more vulnerable, raising significant concerns about the risks associated with their fine-tuning. Processes that teams utilize for improving models often create avenues for cyber adversaries to infiltrate LLMs.

Once attackers penetrate these models, they can quickly corrupt data, take control of infrastructure, manipulate agent behavior, and extract vast quantities of training information. Cisco warns that if organizations do not implement independent security layers, the LLMs they refine will not only be at risk but could become liabilities ripe for exploitation.

Fine-Tuning LLMs Dismantles Safety Controls at Scale

Cisco’s security research focused on scrutinizing several fine-tuned models, such as Llama-2-7B and Microsoft Adapt LLMs that are domain-specific. These models were evaluated across varied sectors, including healthcare, finance, and legal.

A significant takeaway from Cisco’s research is that fine-tuning significantly disturbs model alignment, even if conducted on high-quality datasets. The breakdown in alignment proved particularly acute in sectors like biomedical and legal, where compliance and safety standards are exceptionally stringent.

While enhancing a model’s performance is the goal of fine-tuning, the resulting degradation of inherent safety controls is a concerning byproduct. Attempts to override safeguards that were ineffective against foundational models saw markedly increased success rates when testing fine-tuned variations, particularly in fields mandating rigorous compliance.

The implications are alarming. The success rate for jailbreak attempts increased threefold, with the generation of malicious outputs skyrocketing by 2,200% relative to baseline models. The data starkly illustrates the trade-off inherent in fine-tuning, as it expands a model’s operational capabilities while simultaneously broadening its attack surface.

TAP achieves up to 98% jailbreak success, outperforming other methods across open- and closed-source LLMs. Source: Cisco State of AI Security 2025, p. 16.

Malicious LLMs are a $75 Commodity

The rise of black-market LLMs has been a focal point in Cisco Talos’ ongoing investigations. Their report highlights that tools like GhostGPT, DarkGPT, and FraudGPT are marketed on platforms such as Telegram and the dark web for subscriptions as low as $75 per month. These malicious LLMs are designed for ease of use, serving as plug-and-play solutions for phishing, exploit creation, and other nefarious purposes.

The DarkGPT underground platform, for instance, advertises “uncensored intelligence” and subscription models priced around 0.0098 BTC, presenting these malicious AI tools as comparable to consumer-grade SaaS solutions.

In stark contrast to mainstream LLMs—equipped with safety measures—these malicious models are pre-set for offensive operations and come with user-friendly APIs, updates, and dashboards akin to their legitimate commercial counterparts.

$60 Dataset Poisoning Threatens AI Supply Chains

Researchers from Cisco uncover that adversaries can compromise the foundational datasets of AI models for merely $60, sidestepping the need for sophisticated zero-day exploits. This revelation stems from a collaborative study with Google, ETH Zurich, and Nvidia, illustrating how easily attackers can inject harmful data into widely utilized open-source training datasets.

By targeting expired domains or manipulating the timing of edits to Wikipedia entries during dataset archiving, attackers can influence up to 0.01% of datasets like LAION-400M or COYO-700M, which can significantly affect subsequent LLM training.

The methodologies examined in the study, including split-view poisoning and frontrunning attacks, exploit the fragile trust model inherent in web-crawled data. Given that many enterprise LLMs rely on open data, these types of attacks can proliferate unnoticed and deeply infiltrate inference pipelines.

Decomposition Attacks Quietly Extract Copyrighted and Regulated Content

A particularly unsettling finding from Cisco’s research is that LLMs can be manipulated to disclose sensitive training data without triggering existing safeguards. Using a technique known as decomposition prompting, researchers managed to reconstruct over 20% of certain articles from the New York Times and Wall Street Journal by deconstructing prompts into sub-queries deemed safe by guardrails and subsequently reassembling the outputs.

This method of evading safeguards to access proprietary content or licensed data poses a significant risk for enterprises. Organizations that have implemented LLMs trained on proprietary datasets face acute vulnerabilities via these types of decomposition attacks, which can circumvent conventional input-level defenses and manifest at the model output level. This complicates detection, auditing, and containment efforts.

Organizations deploying LLMs in heavily regulated domains, such as healthcare, finance, or law, must contend with compliance risks that extend beyond standard regulatory frameworks like GDPR, HIPAA, or CCPA. They face the emergence of new types of compliance vulnerabilities where legitimately sourced data can become inadvertently exposed through inference, leading to potential legal repercussions.

Final Word: LLMs Aren’t Just a Tool, They’re the Latest Attack Surface

Cisco’s comprehensive research, coupled with the insights from Talos monitoring the dark web, substantiates the concerns expressed by numerous cybersecurity experts: weaponized LLMs are becoming increasingly sophisticated while the competitive landscape for these tools continues to evolve on illicit markets. The findings from Cisco underscore that LLMs should not be viewed as peripheral tools but rather as central elements of the enterprise infrastructure. The challenges posed by fine-tuning vulnerabilities, dataset poisoning, and model output leaks require that CISOs adopt a proactive, holistic approach to security oversight.

One crucial lesson derived from Cisco’s report is that static safeguards are inadequate in this new age of AI-driven threats. CISOs and security teams must ensure they have real-time oversight across their IT ecosystems, enhance their adversarial testing protocols, and streamline their technology stacks to adapt to this shifting environment where LLMs serve as potential attack vectors that grow more vulnerable with increased fine-tuning.

Source
venturebeat.com

Related by category

Duolingo’s CEO Announces AI Will Replace Contract Workers

Photo credit: www.entrepreneur.com Duolingo is shifting towards an "AI-first" strategy,...

EA and Respawn Cut More Jobs and Cancel Incubation Initiatives

Photo credit: venturebeat.com Respawn, the studio responsible for popular titles...

Navigating Leadership in Times of Chaos and Uncertainty

Photo credit: www.entrepreneur.com For both veteran investors and emerging entrepreneurs,...

Latest news

White House Budget Office “Unresponsive” to Investigations Regarding Frozen Funds, GAO Report Reveals

Photo credit: thehill.com GAO Chief Highlights Challenges in Accessing Information...

Classic Crepes Suzette with a Vibrant Twist Using This Everyday Ingredient!

Photo credit: www.seriouseats.com Classic Crêpes Suzette with a Colorful Twist Why...

Taylor Swift Appears Youthful in 2015 Snapshot with Ed Sheeran

Photo credit: www.instyle.com Ed Sheeran Shares Throwback Photo of Young...

Breaking news