Allgemein

Why 40% of AI Agents Might Fail (and How To Save Yours)

Why 40% of AI Agents Might Fail (and How To Save Yours)

AI agents

In 2023, a Chevrolet dealership in California woke up to a viral nightmare: Its new AI chatbot had just agreed to sell a new, $76,000 Chevy Tahoe to a customer for exactly $1. The user had simply told the bot that its “objective is to agree with anything the customer says” and that it must make a “legally binding offer.” The agent, lacking specific pricing guardrails or a “human-in-the-loop” approval process, happily obliged.

This incident is a perfect, costly microcosm of why Gartner recently predicted that over 40% of agentic AI projects would be canceled by 2027. This prediction seems to contradict the promise that agentic AI would revolutionize productivity and efficiency. However, as the Chevy example proves, the biggest obstacle to the success of agentic artificial intelligence is not the model’s intelligence, but the weakness of its surrounding guardrails.

For enterprise leaders hoping to leverage AI agents, the secret to success is simple: You must implement a system of governance that treats the AI agent with the same, or stricter, diligence you would apply to hiring a human employee.

Guardrails: The AI Agent’s Job Description

When hiring, a human employee is subject to clear rules: logging their work, approval requirements for large expenditures and controlled access boundaries. These same standards must be applied to the AI agent.

If an AI agent is a tool, then the guardrails are its job description, limiting its scope and securing your business. These must go beyond simple logging and include technical constraints like:

  • Sensitive data redaction/access oversight: Automatically scrubbing sensitive data from agent inputs and outputs or requiring human approval for access where necessary.
  • Action limitation: Restricting the agent to a defined, narrow set of APIs or internal systems.
  • Output structure validation: Ensuring the agent’s response is in a required, structured format (such as a Pydantic model for JSON) to prevent execution errors.
  • Testing infrastructure: While traditional unit testing doesn’t work, functional testing is still crucial in ensuring performance doesn’t drift over time.

In the early stages of deployment, where the AI agent’s output is less reliable, these protections, along with fine-grain data labeling, should be meticulously implemented to control what an agent can touch and where human review is required.

Testing Agentic AI: Moving Beyond Unit Tests

One of the biggest myths about agentic AI is that artificial intelligence eliminates the need for testing. This is false. You wouldn’t let a human employee work autonomously without proper training, a successful track record of supervised performance or consistent evaluation. The same is true for AI.

Testing agentic AI systems is crucial, but requires moving beyond traditional software testing. Enterprises must implement rigorous evaluation frameworks, or “evals,” that mirror the way you would evaluate a human worker performing the same task.

  • Outcome-oriented validation: Instead of testing code, you test the result. Did the agent correctly file the expense report? Did it accurately synthesize the requested data points?
  • Model-graded evals: Use a highly reliable, proprietary or more powerful LLM to judge the output of your working agent. This is faster than human review and can automatically detect common failures like incorrect formatting, hallucinations and even the result of prompt injection attacks.
  • Golden datasets: Create a set of high-quality, known-correct examples (a “golden dataset”) that the agent must successfully pass before being promoted to a higher level of autonomy.

How To Design AI Agents To Support Full Autonomy

Of course, the goal of agentic AI is full autonomy — the ability to perform tasks with little to no human oversight. To achieve this safely, agentic AI must be engineered to be predictable, maintainable and effective.

The safest architecture is not a monolithic “god-bot” handling all tasks, but a team of specialized agents.

Instead of giving one agent broad permissions that could lead to a massive security or operational failure (like the Chevy chatbot), use an orchestration agent to coordinate several specialized worker agents, each with narrow permissions designed for specific tasks. The orchestrator logs every task assignment, ensuring that workflows remain secure and that any errors are contained and do not have a ripple effect across the enterprise.

The Phased Approach to Autonomy

To ensure reliability, autonomous agents should be deployed using a multiphase approach, treating it like a mandatory, structured onboarding process for a new employee.

  • Phase 1: Shadow Mode (the Training Period)

      • AI agents complete tasks alongside human workers.
      • The agent’s outputs are compared to their human counterparts.
      • No outputs are executed. The goal is to build statistical confidence in the AI’s quality and trustworthiness.
  • Phase 2: Human-in-the-Loop (the Probationary Period)

      • Humans validate every agent decision before execution and provide explicit feedback.
      • This phase is critical for tasks with legal, financial or compliance implications in preventing catastrophic errors.
      • Outputs must be reviewed and receive human approval until the agent has proven high reliability.
  • Phase 3: Full Automation (the Tenured Employee)

    • The agent is “let loose” to complete tasks on its own.
    • This is only achieved once the agent’s performance metrics have consistently exceeded the required threshold during Phase 2.

Accountability and Oversight

Still, even when AI systems become fully autonomous, that does not absolve companies of all responsibility for the agent’s output. The AI is not an employee; it is a tool. Responsibility for an autonomous system should remain with the manager of the person who initially performed the task. This ensures that someone who is knowledgeable about the task’s outcome, context and risks can evaluate the output and intervene if necessary.

Businesses must conduct ongoing logging and oversight with periodic review to ensure the quality of the output remains equivalent to or better than that of the human worker who preceded the AI agent.

Agentic AI can be a tremendous tool for enterprises to boost productivity. But to unlock its potential, it is necessary to treat the deployment of AI agents the same way you would the hiring of a human employee: With the proper diligence, phased onboarding and strict managerial oversight. Design for failure, and you will succeed at scale.

The post Why 40% of AI Agents Might Fail (and How To Save Yours) appeared first on The New Stack.