Building Autonomous AI Agents That Actually Do Work

Published: (March 7, 2026 at 07:16 PM EST)
6 min read
Source: Dev.to

Source: Dev.to

Niko Alho

The Problem With “AI‑Powered” Tools

Every SaaS product now slaps “AI‑powered” on its landing page, but 99 % of them are just a thin UI around a single LLM call. You still have to:

  1. Decide what to do
  2. Write the prompt
  3. Review the output
  4. Decide what to do next
  5. Repeat forever

That isn’t automation – it’s you doing all the thinking while a machine does the typing.

What an Agent Actually Looks Like

An autonomous agent follows a loop, not a single prompt‑response. The pattern I use is based on the ReAct framework:

┌─────────────────────────────────────┐
│           AGENT LOOP                │
│                                     │
│  ┌──────────┐                       │
│  │ PERCEIVE │◄── APIs, databases,   │
│  └────┬─────┘    live data sources  │
│       │                             │
│  ┌────▼─────┐                       │
│  │  REASON  │◄── LLM + context      │
│  └────┬─────┘    from RAG store     │
│       │                             │
│  ┌────▼─────┐                       │
│  │   ACT    │──► API calls, DB      │
│  └────┬─────┘    writes, alerts     │
│       │                             │
│       └──────────► loop back ◄──────│
└─────────────────────────────────────┘

Each iteration the agent:

  1. Perceives its environment (APIs, databases, live data).
  2. Reasons using an LLM plus any retrieved context.
  3. Acts by calling APIs, writing to a DB, sending alerts, etc.

The key insight: the agent decides when to stop, not you.

The Architecture: Multi‑Agent, Not Monolithic

A single “big” agent quickly hallucinates and loses context. Instead, I split responsibilities across specialized sub‑agents.

from dataclasses import dataclass
from enum import Enum

class AgentRole(Enum):
    RESEARCHER = "researcher"
    AUDITOR    = "auditor"
    STRATEGIST = "strategist"
    EXECUTOR   = "executor"

@dataclass
class AgentTask:
    role: AgentRole
    objective: str
    constraints: list[str]
    context: dict

class AgentOrchestrator:
    def __init__(self, agents: dict[AgentRole, "Agent"]):
        self.agents = agents
        self.shared_state: dict = {}

    async def run_pipeline(self, trigger: dict):
        # 1️⃣ Researcher gathers data
        research = await self.agents[AgentRole.RESEARCHER].execute(
            AgentTask(
                role=AgentRole.RESEARCHER,
                objective="Analyze SERP changes for target keywords",
                constraints=["Use DataForSEO API", "Max 500 queries"],
                context=trigger,
            )
        )

        # 2️⃣ Auditor validates against guidelines
        audit = await self.agents[AgentRole.AUDITOR].execute(
            AgentTask(
                role=AgentRole.AUDITOR,
                objective="Check findings against brand guidelines",
                constraints=["Flag confidence  0.8:
            await self.agents[AgentRole.EXECUTOR].execute(
                AgentTask(
                    role=AgentRole.EXECUTOR,
                    objective="Implement approved changes",
                    constraints=["Dry‑run first", "Log all mutations"],
                    context={"strategy": strategy},
                )
            )

Each sub‑agent has a narrow scope:

  • The Researcher never writes content.
  • The Executor never decides strategy.

This separation keeps hallucinations contained.

The ReAct Loop Inside Each Sub‑Agent

Every sub‑agent runs its own reasoning loop using the ReAct pattern – generating a Thought before each Action.

import json
import openai
from typing import Callable

class Agent:
    def __init__(self, role: AgentRole, tools: list[Callable]):
        self.role = role
        self.tools = {t.__name__: t for t in tools}
        self.client = openai.AsyncOpenAI()

    async def execute(self, task: AgentTask, max_steps: int = 10):
        messages = [
            {"role": "system", "content": self._build_system_prompt(task)},
        ]

        for step in range(max_steps):
            response = await self.client.chat.completions.create(
                model="gpt-4o",
                messages=messages,
                tools=self._tool_schemas(),
            )
            message = response.choices[0].message

            # If the LLM decides to call a tool, run it and feed the result back
            if message.tool_calls:
                for call in message.tool_calls:
                    tool_name = call.function.name
                    args = json.loads(call.function.arguments)
                    result = await self.tools[tool_name](**args)
                    messages.append({"role": "tool", "content": str(result)})
                continue

            # No tool call → final answer
            messages.append({"role": "assistant", "content": message.content})
            break

        return messages[-1]["content"]

    # -----------------------------------------------------------------
    # Helper methods (implementation details omitted for brevity)
    # -----------------------------------------------------------------
    def _build_system_prompt(self, task: AgentTask) -> str:
        return (
            f"You are a {task.role.value} agent.\n"
            f"Objective: {task.objective}\n"
            f"Constraints: {', '.join(task.constraints)}\n"
            f"Context: {json.dumps(task.context)}"
        )

    def _tool_schemas(self) -> list[dict]:
        """Return OpenAI‑compatible tool specifications for self.tools."""
        # Placeholder – in a real implementation you would generate JSON schema
        # objects for each callable in self.tools.
        return []

Each agent:

  1. Receives a system prompt that defines its role, objective, constraints, and context.
  2. Generates a thought (the LLM’s next message).
  3. If the thought includes a tool call, the tool is executed, and the result is fed back as a tool message.
  4. The loop repeats until the LLM produces a final answer (no tool call).

Takeaways

  • Loop‑based agents (perceive → reason → act) are far more powerful than single‑prompt UIs.
  • Specialized sub‑agents keep the overall system grounded and reduce hallucinations.
  • The ReAct pattern gives each agent a disciplined reasoning cycle, allowing it to decide when to call external tools and when to stop.

With this architecture you can let an autonomous SEO assistant run overnight, continuously monitoring SERPs, auditing brand compliance, devising strategies, and executing changes—all without human intervention. 🚀

Example Workflow Snippet

StepReasonAct
while steps 15 % while impressions stayed flat.For each flagged page, fetch the current SERP. Ask the LLM: “Has the search intent shifted? Are competitors using a different content format?”Generate a refresh brief with specific recommendations. If confidence is high enough, push title‑tag updates directly via the WordPress API.

The whole thing runs every 24 hours. I get a Slack notification with what it found and what it did. Most days it finds nothing; some days it catches a decay pattern weeks before I’d have noticed manually.

What Makes This Different From a Cron Script

A cron script does the same thing every time. An agent reasons about what it sees.

  • When the SERP shifts from listicles to video carousels, a cron script keeps optimizing listicles.
  • An agent notices the format change and adjusts its recommendations.

The distinction matters: you define objectives and constraints, not step‑by‑step instructions. The agent figures out the steps.

Guardrails That Actually Matter

Autonomy without guardrails is a liability. Here’s what I enforce:

  • Dry‑run mode – Every mutation gets logged before execution. The agent proposes; a human (or a second agent) approves.
  • Confidence thresholds – Actions below 0.7 confidence get queued for human review instead of auto‑executing.
  • Audit trail – Every thought, tool call, and result is logged. No black boxes.
  • Budget caps – Max API calls per run, max tokens per agent, max mutations per cycle.

The goal isn’t “human out of the loop” — it’s “human on the loop.” You set the boundaries, the agent operates within them, and you review the dashboard.

Getting Started

You don’t need a complex multi‑agent system on day one. Start with a single agent that does one thing:

  1. Pick one repetitive task you do weekly.
  2. Write a Python script that does the perceive step (fetch data).
  3. Add an LLM call for the reason step (analyze the data).
  4. Add an API call for the act step (do something with the analysis).
  5. Wrap it in a loop with a termination condition.

That’s an agent. Everything else is optimization.

Further Reading

Deep dive with full architecture details:
Agentic AI Workflows: Beyond Basic Content Generation

0 views
Back to Blog

Related posts

Read more »