All that glitters is not agentic gold
Investors, vendors and enterprise buyers are racing into this market at full throttle. MarketsandMarkets projects that agentic AI will grow from $13.8 billion in 2025 to nearly $141 billion by 2032.Gartner calls much of what it is seeing “agent-washing,” according to a report released last week. It estimated that out of the thousands of agentic AI vendors, only 130 are "real." According to Gartner, an agentic AI system is defined by its ability to operate with goal-directed autonomy. It must plan, act, and adapt in real time without human micromanagement. The problem is that most tools being sold today don’t come close, it argued.Gartner said some vendors are effectively dressing up scripted bots with a polished interface and marketing them as intelligent agents. The report warns that companies are slapping the “agentic AI” label on everything from old RPA scripts to glorified macros.“Most agentic AI projects right now are early stage experiments or proof of concepts that are mostly driven by hype and are often misapplied,” said Anushree Verma, senior director analyst, Gartner. “This can blind organizations to the real cost and complexity of deploying AI agents at scale, stalling projects from moving into production. They need to cut through the hype to make careful, strategic decisions about where and how they apply this emerging technology.”Meanwhile, enterprise buyers are sold the idea that these tools can make intelligent, autonomous decisions, when in reality many AI bots still can’t even shepherd a help desk ticket along without human intervention.For that reason, Gartner estimated that 40% of agentic AI projects will be canceled by 2027 due to implementation failures and inflated expectations. “Agentic AI projects are being driven by hype, not value,” Verma said.Big dreams, bigger holes
MarketsandMarkets expected agentic AI adoption to expand fastest in IT service management and incident response because these workflows are high-volume, rules-based, and easy to automate. That makes them low-hanging fruit for AI experimentation and high-risk territory for anything that fails silently, critics warn.Flashpoint and MarketsandMarkets both reported that enterprises are already integrating agentic AI into customer service workflows, decision support tools, and infrastructure automation."Amid rising pressure to 'use AI,' defenders are navigating a maze of assumptions, marketing promises, and misconceptions. The technology is moving fast, but so is the confusion around what it can (and can’t) do," Flashpoint said.But according to Cobalt, these systems are often deployed without visibility into how decisions are made or validated. As Cobalt put it: “Visibility into how LLMs make decisions — and how those decisions could be exploited — is still largely missing from enterprise deployments.”Cobalt’s 2025 State of LLM Application Security report found that 32% of tested LLM applications had serious security flaws, and only 21% of the flaws were remediated. The most common issues included prompt injection, model denial-of-service, and data leakage vulnerabilities.GenAI flaws are fixed much less often than other types of flaws, such as API flaws, which are resolved more than 75% of the time, and cloud vulnerabilities, which are fixed in 68% of cases, cited SC Media reporting on the Cobalt report.Developers “building in the dark” means without the security tooling or best practices to anticipate emergent behavior, according to Cobalt. One example highlighted by Cobalt is a healthcare chatbot that it said leaked sensitive patient data after being manipulated through prompt injection. This was caught only during manual human testing.Criminal creativity outpaces enterprise caution
While defenders are still puzzling over governance models and safe deployment, attackers are improvising with jail broken and fine-tuned LLMs to scale fraud, phishing and malware development.In a report released last week, Cisco Talos found that black-market tools such as WormGPT and FraudGPT are built on stripped-down versions of open-source models including LLaMA and GPT-J. These systems are repackaged to generate malicious code, write persuasive phishing emails and guide attackers in evading security measures.Repackaging open-source models typically involves removing safeguards, retraining them on malicious data, or bundling them into plug-and-play tools on dark web forums and Telegram.And the attacks are getting more advanced. Prompt injection attacks, where malicious inputs trick the model into acting outside intended parameters, have gone mainstream. Cisco called these Retrieval Augmented Generation (RAG) pipelines.LLMs using RAG fetch real-time information from external sources to enhance their responses. For instance, if you ask about the weather on a specific day, the model queries a website to retrieve the latest forecast. However, if an attacker gains access to the data source, they could tamper with the information and alter the weather report or embedding hidden instructions to change the model’s response. Such manipulation could mislead users or even target individuals with customized misinformation. Cisco said prompt injection and RAG attacks aren't just a novelty attacks, they have become operationalized. "The threat surface is expanding faster than the defensive playbook,” Cisco said.While these scenarios are less about agentic AI washing, they do play into the larger AI gold-rush narrative impacting enterprise and shadow AI threats security teams must contend with.Defenders, meanwhile, are being asked to both adopt and secure these tools. This leads to what many call a “hype fog,” where decision-makers struggle to separate innovation and unsubstantiated buzz from risk. The term is meant to connote a billboard shrouded in dense fog — message visible, but details obscured.AI manipulation bazaar
In its latest threat intelligence research, Flashpoint chronicled the rise of deepfake-as-a-service marketplaces, fraud-focused LLMs for sale on the dark web, and purpose-built tools to automate identity theft, impersonation, and misinformation. One deepfake-as-a-service kit highlighted by the firm specialized in "custom face generation," voice impersonation and synthetic video."These offerings are designed to fool verification systems used by financial institutions and other regulated industries," Flashpoint said.Flashpoint's approach to integrating AI into its platform decidedly in partnership with "human expertise." Defenders aren’t helpless, only overwhelmed, it noted."Transparency, oversight, and expert interpretation aren’t optional; they’re built into our design.Because in critical missions, AI needs to empower people, not distract them," it maintained. Flashpoint doesn’t promote autonomous AI defenses, rather a fusion of machine-scale monitoring with human analyst insight.Flashpoint's down-to-earth antidote for fog-hype complimented Gartner’s warning over agent-washing where the current marketplace appears to value buzzwords over functionality. Both suggested the disconnect between promises versus reality makes it easier for bad actors to thrive and harder for CISOs to evaluate real value.