Do you actually need an AI Agent? A decision-maker's guide to 3 levels of agentic automation

Founders and operators keep hearing "AI agents" and assume the next step is a complex, developer-led build. In reality, a lot of "agentic" value is already available out of the box — and building too early can be the most expensive way to learn what you actually needed.

Founders and operators keep hearing "AI agents" and assume the next step is a complex, developer-led build.

In reality, a lot of "agentic" value is already available out of the box (research, synthesis, even basic web actions) — and building too early can be the most expensive way to learn what you actually needed.

This post gives you a practical, business-first way to decide between:

  1. off‑the‑shelf agentic capability
  2. configurable / no‑code agent workflows
  3. custom / developer-led agent systems

First: what decision-makers should mean by "AI agent"

A useful definition:

An "agent" is a system that can plan, use tools (search, files, apps, APIs), and take steps toward an outcome — not just generate text.

That's why tools like ChatGPT's Deep Research (multi-step web research with a documented report) feel "agentic," even if you didn't build anything. (help.openai.com)

And why "agent mode" in ChatGPT is even more agentic: it can browse, work with files, connect read-only data sources, and take actions (e.g., fill forms, edit spreadsheets) while keeping you in control. (help.openai.com)

The 3 levels of agentic automation (and when each wins)

A quick comparison (business view)

Level What you're really buying Best for Typical outcome
Level 1: Off‑the‑shelf agentic capability "Instant capability" (no build) One-off or occasional work, fast learning Answers, reports, drafts, plans
Level 2: Configurable / no‑code agent workflows Repeatability + guardrails without a full dev project Standard workflows with clear steps Consistent outputs + light system actions
Level 3: Custom / developer-led agent systems Deep integration, reliability, governance High volume, high stakes, complex systems Real automation inside your product/process

Level 1 — Off-the-shelf: "Don't build. Just get the outcome."

Use this when the job is primarily thinking and synthesis, not integration-heavy automation.

Examples of Level 1 "agentic" tools:

  • OpenAI Deep Research: multi-step web research that produces a cited report; can use web + uploaded files, and can use connected sources you enable. (help.openai.com)
  • Perplexity Research mode: runs many searches, reads many sources, and produces a comprehensive report. (perplexity.ai)
  • Gemini Deep Research: browses the web and (optionally) uses Google workspace context like Gmail/Drive/Chat to create multi-page reports. (gemini.google)
  • General-purpose "do tasks for me" agents (example: Manus has been described publicly as an autonomous agent capable of tasks like coding and data analysis). (businessinsider.com)

When Level 1 is enough

  • You're still figuring out the process
  • It's low frequency
  • You need a research brief, plan, draft, or options — not "push buttons in 8 systems"

Decision-maker KPI: Speed-to-insight (hours saved, clarity gained, better decisions).

Level 2 — Configurable / no-code: "Make it repeatable before you make it custom."

This is where many teams should land first: you've proven the workflow is valuable, and now you want it more consistent.

Two common "Level 2" approaches:

A) Configure a "no-code agent" in a consumer/workspace product

Example: Custom GPTs can include instructions, "knowledge" files, and actions that connect to external APIs. (platform.openai.com)

Important reality check on proprietary data

  • The "Knowledge in GPTs" feature allows attaching up to 20 files, each up to 512 MB (and up to 2,000,000 tokens per file). That's great for handbooks and playbooks — not for "our entire data warehouse." (help.openai.com)

B) Use a no/low-code workflow builder designed for agent workflows

Example: OpenAI's Agent Builder is a visual canvas for multi-step agent workflows you can debug and then embed (e.g., via ChatKit) or export into code. (platform.openai.com)

Decision-maker KPI: Consistency (same inputs → same structured output; fewer errors; predictable handling).

Level 3 — Developer-led: "When the agent must actually run part of the business."

Choose Level 3 when:

  • the workflow is high volume (thousands of runs)
  • the workflow is high stakes (money, compliance, safety, reputation)
  • the workflow needs deep integration (multiple internal systems, permissions, audit trails)
  • you need custom UX and operational controls

Developer-led orchestration frameworks exist because real-world work needs: state, retries, human approvals, observability, and multi-step logic.

Examples:

  • OpenAI Agents SDK: supports tools, handoffs to specialized agents, streaming, and "a full trace of what happened." (platform.openai.com)
  • Google Agent Development Kit (ADK): open-source framework; model-agnostic; designed to help build/deploy/orchestrate agent architectures; Google recommends deploying to a managed runtime in Vertex AI Agent Engine. (docs.cloud.google.com)
  • LangGraph: orchestration framework for long-running/stateful agents with durable execution and human-in-the-loop. (github.com)
  • CrewAI: multi-agent systems with guardrails, memory/knowledge, and observability concepts. (docs.crewai.com)

Decision-maker KPI: Throughput + control (automation that's auditable, safer, and integrated).

The "should we build?" rubric (use this before you scope anything)

Ask these six questions:

  1. How often does this happen? (daily vs quarterly)
  2. What's the cost of a mistake? (annoying vs legally/financially risky)
  3. Does the system need to take actions in other tools? (or just produce info?)
  4. Is the workflow stable and documented? (or still evolving?)
  5. What data boundaries exist? (PII, regulated data, customer trust)
  6. Do you need a real UX and audit trail? (approvals, logs, reporting)

Your answers almost always map cleanly to Level 1 / 2 / 3.

A note on safety (why "agentic" raises the stakes)

When an agent reads untrusted text (web pages, emails, documents) and can call tools, you increase the risk of prompt injection (malicious text trying to override instructions) and private data leakage. (platform.openai.com)

That doesn't mean "don't do it." It means: don't skip the decision framework — and don't jump to full autonomy when an approval step is enough.

What to do next (practical move)

If you're exploring agents for your business, your fastest path is:

  • Run Level 1 on a real workflow for a week (prove value)
  • Move to Level 2 to standardize and reduce variability
  • Only then justify Level 3 (integration + reliability + governance)

CTA: If you want, I offer a short "Agent Opportunity Audit" where we pick 1–2 workflows, score them with the rubric above, and decide what level you actually need.