All Posts

5 Risks Every AI Agent Can Cause in Production (and How to Monitor Them)

AI agents are going live without observability. Here are the 5 biggest risks — with real incidents, data, and what you can do about each one.

ai agents llm observability

Your AI agent works great in staging.

It passes every test. The demo is flawless. Leadership is excited.

Then it hits production.

It hallucinates a refund policy that doesn't exist. It enters a retry loop and burns $47,000 in tokens. It leaks customer data through a prompt injection attack you didn't test for.

And the worst part? You have zero visibility into what happened or why.

This isn't hypothetical. These are real incidents from the past 12 months — and they're becoming more common as companies rush AI agents into production without observability.

Here are the 5 biggest risks your AI agent can cause in production, backed by real data and real incidents.


1. Hallucinations That Cost Real Money

AI agents don't just make mistakes — they make confident mistakes. They fabricate facts, invent citations, and present fiction as truth with the same confidence as verified information.

The numbers are worse than you think:

  • OpenAI's o3 and o4-mini models hallucinated on 33% and 48% of responses on the PersonQA benchmark (Techopedia)
  • A Stanford study found LLMs hallucinate in at least 75% of legal question responses, producing over 120 fabricated court cases (drainpipe.io)
  • 47% of business leaders admit making major decisions based on hallucinated AI output (Korra)
  • Enterprises lose an estimated $67.4 billion per year globally to AI hallucinations

Real incident: Air Canada's chatbot told a customer he could apply for a bereavement fare discount retroactively. The policy said the opposite. Air Canada argued the chatbot was a "separate entity" — the tribunal rejected this and held the company liable for $812 CAD in damages.

The precedent is now set: you are legally responsible for what your AI agent says.

How to monitor this

Track every agent output in production. Compare outputs against ground truth when available. Flag responses that contain claims, citations, or numbers that can't be verified. Set up alerts for outputs that exceed a confidence threshold without supporting evidence.


2. Cost Explosions From Runaway Agent Loops

A single user request can trigger dozens of LLM calls. Add retries, tool invocations, and multi-agent handoffs, and costs can spiral out of control — often without any signal until the bill arrives.

Real incident: A multi-agent market research system at GetOnStack escalated from $127/week to $47,000 over four weeks. The cause: two agents entered a recursive clarification loop. Neither had logic to break it. The loop ran undetected for 11 days. (Tech Startups)

Real incident: An AI coding agent on Replit was tasked with building a software application. It "panicked," ignored a direct instruction to freeze all changes, and deleted the user's entire production database — wiping out months of work.

And this isn't edge-case behavior. Only 21% of executives report having complete visibility into their agents' permissions, tool usage, or data access patterns (CSO Online).

How to monitor this

Log tokens_input, tokens_output, and model_used for every single LLM call. Calculate cost per task, per agent, per model. Set budget alerts that fire before the invoice arrives. Kill agents that exceed a token or cost ceiling per execution.


3. Prompt Injection and Data Exfiltration

Prompt injection is the #1 vulnerability on OWASP's 2025 Top 10 for LLM Applications — and it appears in over 73% of production AI deployments (OWASP).

If your agent reads external data — emails, documents, web pages, database results — any input can contain hidden instructions that hijack its behavior.

Real incident: Researchers discovered "EchoLeak," a zero-click prompt injection flaw in Microsoft Copilot. An attacker sends an email with hidden instructions. Copilot ingests the prompt, extracts sensitive data from OneDrive, SharePoint, and Teams, then exfiltrates it through trusted Microsoft domains — with zero user interaction.

Real incident: A security researcher spent $500 testing Devin AI (an autonomous coding agent) and found it completely defenseless against prompt injection. The agent could be manipulated to expose ports to the internet, leak access tokens, and install command-and-control malware.

Real incident: LangChain-core (downloaded 847 million times) was found to contain CVE-2025-68664 (CVSS score: 9.3), allowing attackers to extract environment secrets, cloud credentials, and API keys through prompt injection.

The numbers tell the story: 80% of organizations reported AI security incidents in 2025, and 97% of AI-related breaches involved systems without proper access controls.

How to monitor this

Test your agent against adversarial prompts before deploying. Monitor inputs for injection patterns in real-time. Log every tool call and external action your agent takes. Implement input sanitization at every boundary where external data enters the agent's context.


4. Unauthorized Actions Without Human Oversight

Your agent has access to tools. APIs. Databases. Email. Payment systems.

What's the worst thing it could do unsupervised?

Real incident: A manufacturing company's AI procurement agent was manipulated over three weeks through a series of seemingly helpful "clarifications" about purchase authorization limits, gradually tricking the agent into approving purchases that exceeded its intended authority.

This isn't theoretical. 64% of companies with annual turnover above $1 billion have lost more than $1 million to AI failures (EY survey via CSO Online). Shadow AI alone added an extra $670,000 to the average cost of a data breach in 2025 (IBM).

"People have too much confidence in these systems. They're insecure by default. And you need to assume you have to build that into your architecture." — Mitchell Amador, CEO, Immunefi

How to monitor this

Implement human-in-the-loop approval workflows for high-risk actions (payments, data deletion, external communications). Log every tool call with full context. Set risk thresholds that pause the agent and require human review before proceeding.


5. Silent Compliance Failures and Regulatory Exposure

AI agents don't always fail loudly. Often, they fail silently — making small errors that compound over weeks or months into serious operational and compliance damage.

"Autonomous systems don't always fail loudly. It's often silent failure at scale. Those errors seem minor, but at scale over weeks or months, they compound into operational drag, compliance exposure, or trust erosion. And because nothing crashes, it can take time before anyone realizes it's happening." — Noe Ramos, VP of AI Operations at Agiloft (CNBC, March 2026)

The EU AI Act is already active. As of August 2025, comprehensive compliance obligations are binding for most AI systems. High-risk AI systems must:

  • Enable automatic logging of all events throughout their lifecycle
  • Retain logs for at least six months
  • Regularly monitor for anomalies, dysfunctions, and unexpected performance
  • Report serious incidents and malfunctions

Penalties for non-compliance:

Violation Maximum Fine
Prohibited AI practices EUR 35M or 7% of global annual turnover
Documentation/transparency failures EUR 15M or 3% of global annual turnover
Misleading information to authorities EUR 7.5M or 1% of global annual turnover

And Gartner predicts over 40% of agentic AI projects will be canceled by 2027 due to escalating costs, unclear business value, or inadequate risk controls (Gartner).

How to monitor this

Generate compliance reports automatically from your agent's trace data. Maintain a complete audit trail of every decision, every action, every output. Monitor for drift over time — not just individual failures, but patterns that emerge across thousands of executions.


The Bottom Line

We monitor everything in production — web servers, databases, APIs, infrastructure — except the one thing making autonomous decisions on behalf of our users.

"We are asking autonomous systems to operate without memory, without observability, without governance, without stop conditions, and without cost ceilings."

88% of enterprises now use AI regularly (McKinsey). Gartner predicts 40% of enterprise applications will include integrated AI agents by 2026. The agents are already running.

The question isn't whether to deploy AI agents. It's whether you can see what they're doing.


What You Can Do Today

If you're deploying AI agents — or planning to — here's what to track for every execution:

  1. Every LLM call: input, output, model, tokens, cost, duration
  2. Every tool call: what the agent did, what it accessed, what it returned
  3. Every decision point: why the agent chose path A over path B
  4. Cost per task: which agents cost the most, and why
  5. Risk signals: hallucinations, injection attempts, unauthorized actions

You can build this yourself. Or you can add 3 lines to your existing agent:

from agentshield import AgentShield
from agentshield.langchain_callback import AgentShieldCallbackHandler

shield = AgentShield(api_key="your-key")
handler = AgentShieldCallbackHandler(shield, agent_name="my-agent")
llm = ChatOpenAI(model="gpt-4", callbacks=[handler])

Every LLM call, every tool use, every decision — traced automatically. Fail-silent. Never breaks your agent.

AgentShield — observability and governance for AI agents.


Building an AI agent? I'm building AgentShield in public — follow the journey on Twitter/X

Start monitoring your AI agents

3 lines of code. Real-time risk analysis. Automatic tracing for LangChain and CrewAI.