Business, AI, ROI·

Prompt Injection Has Left the Chatbot

What ServiceNow's AI agent incidents reveal about enterprise workflow integrity and the next security failure.

Prompt injection becomes dangerous when it is connected to workflows that can act.

Prompt Injection Has Left the Chatbot

What ServiceNow's AI Agent Incidents Teach Us About the Next Enterprise Security Failure

For years, prompt injection was treated as a curiosity -- a clever way to make a chatbot ignore its rules or say something unexpected. Interesting for demos. Annoying in production. Rarely considered a serious enterprise security risk.

The recent disclosures involving ServiceNow's AI agents should put that misconception to rest.

These incidents do not represent a failure of large language models in isolation. They represent a failure of system design -- one that becomes inevitable when natural-language systems are embedded into trusted enterprise workflows without redefining security boundaries.

Prompt injection is no longer a "model issue." It is now a workflow integrity problem.

Why prompt injection still works -- and always will

In an earlier post, How Prompt Injection Attacks Actually Work, we explained the uncomfortable truth behind these attacks:

Large language models do not reliably distinguish between instructions and data. They treat all text as potentially actionable unless the system explicitly constrains them.

This is not a bug. It is a consequence of how LLMs work.

Prompt injection exploits that property by hiding instructions inside content that appears inert to humans -- ticket descriptions, comments, emails, knowledge articles, synced web pages. The attacker does not need direct access to the model. They only need to ensure the model eventually reads their text.

As long as LLMs are allowed to:

  1. Ingest untrusted text
  2. Interpret it autonomously
  3. Take real actions using tools

Prompt injection remains viable.

ServiceNow demonstrates what happens when all three conditions are met at enterprise scale.

From language confusion to operational impact

Recent reporting and independent research showed that ServiceNow AI agents could, under certain configurations:

  • Be tricked into impersonating users
  • Invoke privileged workflows
  • Recruit other agents to perform restricted actions
  • Modify records, exfiltrate data, or trigger downstream automation

These were not classic software exploits. There was no memory corruption, no malformed packet, no buffer overflow.

Instead, the attack surface was text -- processed by AI agents that were deeply integrated into identity, workflow, and automation systems.

Once an AI agent is trusted to act, every piece of text it reads becomes a potential control input.

The exploit chain most teams are missing

To understand why this matters, it helps to visualize the full attack path. In traditional security reviews, these steps often appear disconnected. In AI-driven systems, they collapse into a single flow.

Visual exploit chain: from text to impact

This is not theoretical. Every box in this chain already exists in modern enterprise platforms.

Second-order prompt injection: the "confused deputy" returns

One of the most concerning aspects of the ServiceNow findings is what researchers call second-order prompt injection.

In these scenarios:

  • A low-privilege agent ingests malicious text
  • That agent delegates or requests help from a higher-privilege agent
  • The higher-privilege agent executes harmful actions believing it is assisting a legitimate workflow

This is the classic confused-deputy problem -- except the "confusion" is expressed in natural language, and the deputy is an AI agent.

From a defender's perspective, this looks less like a prompt attack and more like lateral movement inside an automation system.

Why this is not just a ServiceNow problem

ServiceNow is not unique. It is simply early.

Any platform that embeds AI agents into:

  • Ticketing systems
  • CRM platforms
  • Identity workflows
  • DevOps pipelines
  • Knowledge management
  • Internal copilots

inherits the same risk profile.

The moment an AI agent can do something -- not just answer a question -- text becomes executable input.

Simon Willison has described this broader trend as a form of "normalization of deviance": we gradually grant systems more autonomy because nothing bad has happened yet. The failure only becomes visible once the blast radius is large.

ServiceNow is a preview, not an outlier.

What security teams need to change -- now

The solution is not "better prompts" or "more guardrails." Those help, but they do not eliminate the underlying risk.

The organizations that will avoid serious incidents are already making structural changes:

  1. Remove ambient authority
    AI agents should not inherit broad user permissions by default.
  2. Segment agents by impact, not convenience
    Treat agent teams like network zones. Restrict discovery and delegation.
  3. Require human approval for irreversible actions
    Email, exports, access changes, and cross-system writes should never be fully autonomous.
  4. Treat text fields as an attack surface
    Tickets, comments, and synced content now require threat modeling.
  5. Instrument agents like production services
    Log prompts, tool calls, delegation paths, and identity context -- then alert on deviations.

Most importantly, assume that prompt injection cannot be fully prevented. Design systems so that when it happens, damage is limited, visible, and recoverable.

Final thought

Prompt injection did not suddenly become dangerous. We simply connected it to systems that matter.

ServiceNow's AI agent incidents confirm what many of us have been warning for years: when language controls software, language becomes an attack vector.

The question for every organization now is not whether this risk applies -- but whether they are still modeling AI agents as chatbots, or finally treating them as privileged software components.

Sources & Further Reading

The Data Consulting Company — How Prompt Injection Attacks Actually Work https://www.thedataconsultingcompany.com/blog/how-prompt-injection-attacks-actually-work

Dark Reading — AI Vulnerability in ServiceNow Raises Concerns for Remote Workforce Security https://www.darkreading.com/remote-workforce/ai-vulnerability-servicenow

The Hacker News — ServiceNow AI Agents Can Be Tricked Into Harmful Actions https://thehackernews.com/2025/11/servicenow-ai-agents-can-be-tricked.html

PointGuard AI — ServiceNow AI Agents Can Be Tricked Into Harmful Actions https://www.pointguardai.com/ai-security-incidents/servicenow-ai-agents-can-be-tricked-into-harmful-actions

Simon Willison — Moltbook https://simonwillison.net/2026/Jan/30/moltbook/

OWASP — Prompt Injection https://owasp.org/www-community/attacks/PromptInjection