agentsautonomytool-usereasoningplanning

AI Agents & Tool Use: When AI Stops Predicting and Starts Acting

The shift from passive models to active agents that reason and take action

AI Resources Team··10 min read

For years, AI was reactive. You asked it a question. It generated an answer. Done.

Now it's becoming proactive. You ask it to "find me a flight to Tokyo." It browses the web. Checks multiple airlines. Compares prices. Handles the booking. All automatically.

You ask it to "debug this error in my code." It reads your error message. Runs the code. Checks output. Modifies code. Reruns. Iterates until it's fixed.

This is the shift from "AI that predicts" to "AI that acts." This is agents.


What Is an AI Agent?

An AI agent is a system that:

  1. Perceives the world (sees inputs, data, environment)
  2. Reasons about the situation (what's happening? what does it mean?)
  3. Plans actions (what should I do?)
  4. Acts on those plans (takes actual steps, uses tools)
  5. Observes results (what happened? did it work?)
  6. Iterates (adjusts strategy based on results)

It's like a person, but in software.

A chatbot is not an agent. It responds to your question and stops. An agent continues, takes actions, and adapts.

Example:

You: "I need to hire a senior software engineer. Help me."

Chatbot response: "Here are tips for hiring engineers: post on job boards, use recruiters, interview thoroughly, check references..."

Agent response:

  • Posts a job description to LinkedIn, AngelList, and HackerNews
  • Searches GitHub for engineers matching your criteria
  • Sends outreach emails using your company voice
  • Collects responses
  • Schedules interviews
  • Sends interview prep to candidates
  • Conducts interviews (with human feedback)
  • Checks references
  • Sends offers
  • Does all of this iteratively, asking for your input when needed

The agent does things. The chatbot just talks about them.


Function Calling & Tool Use

The technical foundation of agents is function calling (or tool use).

An LLM can't browse the web or write to a database natively. So you give it access to tools (functions it can call).

Example tools:

  • google_search(query) — search the web
  • read_file(path) — read a file
  • write_code(language, code) — write code
  • execute_command(cmd) — run a command
  • send_email(to, subject, body) — send an email
  • database_query(sql) — query a database

The agent decides when to use which tool.

OpenAI Function Calling

OpenAI's API lets you define functions:

from openai import OpenAI

client = OpenAI()

tools = [
    {
        "type": "function",
        "function": {
            "name": "search_web",
            "description": "Search the web for information",
            "parameters": {
                "type": "object",
                "properties": {
                    "query": {"type": "string", "description": "Search query"}
                },
                "required": ["query"]
            }
        }
    },
    {
        "type": "function",
        "function": {
            "name": "book_flight",
            "description": "Book a flight to a destination",
            "parameters": {
                "type": "object",
                "properties": {
                    "destination": {"type": "string"},
                    "date": {"type": "string"},
                    "airline": {"type": "string"}
                },
                "required": ["destination", "date"]
            }
        }
    }
]

response = client.chat.completions.create(
    model="gpt-4",
    messages=[{"role": "user", "content": "Book me a flight to Tokyo next week"}],
    tools=tools,
    tool_choice="auto"
)

# The model might respond:
# "I'll search for flights to Tokyo next week"
# Then call the search_web tool

The model returns which tool to call and with what parameters. Your code executes it. The result goes back to the model. The model keeps going.

Claude Tool Use

Anthropic's Claude also supports tool use:

from anthropic import Anthropic

client = Anthropic()

tools = [
    {
        "name": "search_web",
        "description": "Search the web",
        "input_schema": {
            "type": "object",
            "properties": {
                "query": {"type": "string"}
            },
            "required": ["query"]
        }
    }
]

response = client.messages.create(
    model="claude-3-5-sonnet-20241022",
    max_tokens=1024,
    tools=tools,
    messages=[{"role": "user", "content": "What's the latest news on AI?"}]
)

Same idea. Different implementation.


Agent Architectures & Patterns

ReAct (Reasoning + Acting)

The most popular pattern. The agent alternates between reasoning and acting.

Thought: I need to find flights to Tokyo
Action: search_web("flights to Tokyo from New York next week")
Observation: [search results with flights]
Thought: I found several options. Let me get prices for the cheapest
Action: get_flight_price("Delta flight 123", "economy")
Observation: [prices]
Thought: Got the prices. I should book the cheapest one
Action: book_flight("Delta 123", "economy")
Observation: [confirmation]
Thought: Done. I booked the flight.

The agent narrates its thinking. This is helpful for debugging and human understanding.

Planning Agents

Some agents plan before acting:

User request: "Help me prepare for a job interview"

Agent planning phase:
1. Research the company
2. Identify likely interview questions
3. Prepare answers
4. Practice interview format
5. Get feedback

Execution phase:
[Executes each step]

Planning first is powerful because it avoids wasted actions. The agent thinks through the whole task before starting.

Hierarchical Agents

Some problems need multiple levels:

High-level agent: "Find me a house in Austin" → Delegates to lower-level agents:

  • Search agent: "Find houses matching criteria"
  • Finance agent: "Get mortgage pre-approval"
  • Legal agent: "Check title and contracts"
  • Inspection agent: "Schedule and manage inspection"

The high-level agent coordinates. Lower-level agents are specialists.


Real Examples of Agents in 2025

Devin: The AI Software Engineer

Cognition Labs released Devin in 2024. It's an AI agent that can write code, fix bugs, and deploy applications.

Example: "Write a Python API that handles user authentication using JWT."

Devin:

  • Writes the code
  • Creates tests
  • Checks for bugs
  • Runs the tests
  • Deploys it
  • Monitors errors

It's not perfect (still needs human oversight), but it's genuinely useful for programming tasks.

AutoGPT

An open-source project that shows agents in action. Give it a goal:

"Increase Twitter followers by 1,000."

AutoGPT:

  • Researches growth strategies
  • Identifies trending topics
  • Drafts tweets
  • Schedules them
  • Monitors engagement
  • Adjusts strategy

Again, not perfect, but demonstrates the agent pattern.

Claude Web Browsing Agent

Anthropic's Claude can browse the web. You can ask it:

"What's the current stock price of Tesla? How has it changed in the last month?"

Claude:

  • Searches for Tesla stock
  • Browses finance websites
  • Gathers historical data
  • Analyzes trends
  • Reports back with sources

Real-World: Insurance Claims

Some insurance companies are deploying agents:

"I need to file a claim for my car accident."

Agent:

  • Asks for details (car, damage, insurance policy)
  • Gathers information
  • Takes photos of damage (via your phone)
  • Generates estimate
  • Files the claim
  • Follows up on status

Humans handle complex cases. Agents handle routine ones. Both get done faster.


ReAct Pattern: Deep Dive

ReAct is so popular it's worth understanding deeply.

The loop:

Question: "Who directed the movie that won Best Picture at the 2024 Oscars?"

[Thought] I need to find which movie won Best Picture at the 2024 Oscars, then find its director.

[Action] search_web("Best Picture Oscar winner 2024")

[Observation] "Oppenheimer" won Best Picture at the 96th Academy Awards in 2024.

[Thought] Now I know the movie is Oppenheimer. I need to find who directed it.

[Action] search_web("Oppenheimer director")

[Observation] Oppenheimer was directed by Christopher Nolan.

[Thought] I have the answer.

[Answer] Christopher Nolan directed Oppenheimer, which won Best Picture at the 2024 Oscars.

The agent:

  1. Thinks out loud (transparency)
  2. Takes an action (calls a tool)
  3. Observes the result
  4. Repeats until it has an answer

This is human-like reasoning. And it's surprisingly effective.


Challenges with Agents

Hallucination & False Confidence

Agents can confidently make up information. An agent might claim to have booked a flight when it actually failed.

Solution: Human oversight for important actions. The agent suggests, humans approve.

Tool Abuse

An agent might call a tool incorrectly or loop forever.

Agent calls delete_all_files() by mistake. Disaster.

Solution: Careful tool design. Permissions. Limits on tool use.

Complexity

Debugging agents is hard. The interaction between reasoning and actions is complex.

Solution: Logging. Explainability. Testing.

Cost

Agents take multiple API calls. A simple question might need 5-10 LLM calls.

ReAct style: 1 call to think, 1 to execute, 1 to reason on result = 3 calls per step. If a task takes 10 steps, that's 30 calls.

At $0.001 per call (rough pricing), a single task costs $0.03. Add up to 10,000 tasks per day = $300/day = $110k/year.

For companies this is manageable. For startups, it stacks up.


Limitations of Current Agents

They're not fully autonomous. Even Devin (the AI engineer) needs human review. It's not "set it and forget it."

They struggle with long chains of reasoning. After 10+ steps, errors compound.

They need clear, actionable tools. If a tool doesn't exist, the agent can't use it.

They're expensive. Multiple API calls add up.

They hallucinate. Even powerful models make up information.


The Future of Agents

2025-2026: Agent Proliferation

Every company will experiment with agents:

  • Customer service agents
  • Code agents
  • Sales agents
  • HR agents
  • Research agents

Most will be hybrid (agent + human). Agents handle routine stuff. Humans handle complex stuff.

Specialized Agents

Instead of one general agent, specialized ones:

  • A web research agent (best at searching/analyzing)
  • A coding agent (best at writing code)
  • A writing agent (best at generating content)

Each optimized for its domain.

Agentic Frameworks

Platforms emerging to build agents:

  • LangChain: Tool use, memory, planning
  • Pydantic AI: Type-safe agent building
  • Anthropic's prompt libraries: Agent patterns

Multi-Agent Systems

Agents working together:

  • Agent 1 searches for information
  • Agent 2 analyzes it
  • Agent 3 writes a report

This is where things get really powerful.


The Open Question: Safety

As agents become more autonomous, safety becomes critical.

An agent that can browse the web and run code could:

  • Delete important files
  • Send harmful emails
  • Leak confidential data
  • Make bad financial decisions

Solutions in development:

  • Sandboxing (agents run in isolated environments)
  • Permissions (agents can only access approved tools)
  • Monitoring (humans watch what agents do)
  • Rollback (if something goes wrong, undo it)

None of this is perfect. This is an ongoing research problem.


FAQs

Q: Can agents replace humans? Not yet. They're tools that augment humans. A human + agent team is better than either alone.

Q: How much does an agent cost? Depends on tool use. A simple agent: $0.01-0.10 per task. Complex ones: $0.50-5.00 per task.

Q: Can I build an agent myself? Yes. Use LangChain, Claude API, or OpenAI API. There are tutorials. It's not trivial but totally doable.

Q: What if my agent makes a mistake? Design it so mistakes are caught by humans. Use approval workflows. Have rollback plans.

Q: Are agents really autonomous? Not really. They're executing a plan with tools. The plan might be generated by an LLM, but it's not truly autonomous decision-making.


The Bottom Line

Agents represent a shift from "AI that answers questions" to "AI that takes action."

This is powerful. It's also risky. An agent that can book flights or write code needs careful oversight.

But the benefits are clear: automation of complex tasks, human time freed up, better consistency.

Expect agents everywhere in 2025-2026. Customer service, coding, research, writing. Every domain.

The best approach: start with human-in-the-loop agents. Agent suggests → human approves → agent acts. As you build trust, add more autonomy.

The future is agentic. Prepare accordingly.


Next up: Artificial General Intelligence (AGI): Are We Close? — Because agents are a step toward AGI. But how many steps?


Keep Learning