Agents, Workflows, and Tools: Navigating Through the Hype

Valtteri KarestoJune 18, 2025
Breaking down the spectrum from fixed workflows to goal-driven agents in AI system design.

The Core Distinction That Matters

Workflows are recipes. Agents are chefs.

We think of workflows as deterministic, graph-based execution paths versus agents as goal-oriented systems with decision-making capabilities. But as you will see, the boundaries blur quickly.

Here's an analogy that might help: Think of workflows like a recipe and agents like a chef.

A recipe (workflow) has:

  • Fixed steps in sequence
  • Predictable outputs given the same inputs
  • No decision-making beyond simple conditionals ("if dough is too dry, add water")
  • Clear start and end points

A chef (agent) has:

  • A goal ("make dinner for 6 people")
  • Access to tools (knives, ovens, recipes)
  • Ability to make decisions ("the market is out of salmon, I'll use chicken instead")
  • Can invoke sub-tasks including following recipes

But here's where it gets interesting - and mirrors your observation about recursive complexity:

  • Tools as Workflows: A chef might use a bread machine (a tool that executes a workflow). The chef decides when to use it, but the machine follows its program.
  • Workflows with Agent Steps: A restaurant might have a workflow where "Step 3: Sous chef prepares sauce" - embedding an agent within a deterministic flow.
  • Agents Using Agents: A head chef might delegate to sous chefs (other agents), who might delegate further.

The key insight is that these aren't really different categories but different levels of autonomy on a spectrum:

Fixed Script → Parameterized Workflow → Conditional Workflow → Goal-Directed Agent → Autonomous Agent

Tool Calling: Where the Magic Actually Happens

Here's the key insight: the agentic behavior you're seeing doesn't come from agent frameworks—it comes from the LLMs themselves.

Modern LLMs like GPT, Claude, and Gemini have native tool-calling capabilities. When you see an "agent" that can search the web, analyze data, and write code, you're witnessing the LLM's inherent ability to reason about and orchestrate tool usage.

Automatic Tool Chaining The real magic happens when LLMs chain tools based on previous results. For "I need to book a flight to the cheapest European city with good weather next week," the LLM will:

  1. Search cheapest European flights → Find Budapest, Prague, Lisbon
  2. Check weather for each → Budapest (rainy), Prague (cloudy), Lisbon (sunny)
  3. Focus on Lisbon based on weather
  4. Search specific Lisbon flights
  5. Present recommendation combining price and weather data

This is genuine agentic behavior: goal-oriented reasoning, dynamic planning, and adaptive execution. But it's the LLM's native capability, not framework innovation.

What Frameworks Actually Provide Most "agent frameworks" essentially offer:

  • Tool Integration: Simplified APIs for connecting business systems
  • Conversation/State/History Management: Handling multi-turn interactions and context
  • Error Handling: Managing failed operations and retries
  • UI Components: Pre-built chat interfaces and dashboards

The intelligence isn't in the framework—it's in the LLM's ability to reason about tool usage. Understanding this has important implications for your build vs. buy decisions and vendor independence.

Agents as Abstraction Layers

Ultimately, agents are abstraction layers that package LLM reasoning, tool orchestration, and workflow management into reusable components—valuable for operational consistency, but the core intelligence still comes from the underlying LLM.

The Value of Prompt Transparency

Many frameworks provide a "black box" approach—you define roles, goals, and tools, but the actual prompts sent to the LLM are hidden. While this offers excellent prompt engineering out of the box, you lose transparency and it becomes difficult to reverse engineer what actually got sent to the model when debugging or optimizing performance.

For example, in OpenAI's Agents SDK and Google's ADK, the popular "handoff" feature that routes conversations between specialist agents is simply a tool call with built-in system prompts that you can figure out only by reading the code.

By treating prompts as first-class code (inspired by the 12-Factor Agents methodology), we maintain full control over what's sent to the LLM and can optimize precisely for our specific use cases.

Summary: Navigating Through the Hype

The AI industry is full of marketing buzzwords, but understanding the fundamentals helps you make better strategic decisions:

The Reality Check: Whether you call it an agent, workflow, or tool—the intelligence comes from the LLM itself. Modern LLMs like GPT-4 and Claude have native tool-calling capabilities that enable sophisticated, goal-oriented behavior without any framework magic.

The Spectrum: Rather than distinct categories, think of a spectrum from fixed scripts to autonomous agents. Most practical AI applications combine deterministic workflows with LLM decision-making at key points.

The Trade-off: Agent frameworks offer operational convenience but reduce transparency. The choice isn't whether frameworks are good or bad—it's whether their benefits outweigh the loss of control for your specific use case.

The Strategic Insight: Your competitive advantage doesn't come from the agent wrapper—it comes from your domain expertise, proprietary data, and the quality of your prompts and tools. The LLM provides the intelligence; everything else is implementation detail.

Understanding these fundamentals helps you evaluate vendors more effectively, make better build-vs-buy decisions, and avoid paying enterprise premiums for capabilities you could integrate directly.

Continue reading

Thoughts on the Model Context Protocol Part 2

In the previous post we introduced some concerns around Model Context Protocol, things have updated quite a bit, here's a quick review.

Valtteri Karesto

Valtteri Karesto

CTO, Founder

Thoughts on the Model Context Protocol

Model Context Protocol (MCP) is a new open protocol that aims to standardize how AI applications connect to custom data sources (files, photos) and tools (eg. functions that can fetch inform from third party systems). It was released by Anthropic, but in theory, any AI application could support it. At the moment, Claude Desktop and Cursor are among the most popular applications that support MCP. This means that third party developers can build custom tools and other capabilities host applications like Claude Desktop can use.

Valtteri Karesto

Valtteri Karesto

CTO, Founder

Synthetic Data in 2025: Revolutionizing GenAI Model Performance

How Synthetic Data is Powering the Next Generation of Efficient, Specialized AI Models

Valtteri Karesto

Valtteri Karesto

CTO, Founder

Optimizing current business vs. unlocking new business opportunities using GenAI

Transformative impact of Generative AI, particularly Large Language Models, on business optimization and the creation of new opportunities, highlighting its applications in various industries like real estate and the importance of exploring beyond current business models to fully leverage AI's potential

Tuomas Martin

Tuomas Martin

VP of Sales

Founder Conversations: A Week of LLM Insights in the Bay Area

We spent a week talking to founders and builders at Ray Summit, TechCrunch Disrupt, and various Bay Area GenAI meetups to understand the challenges they face when building value-providing LLM-based apps.

Joni Juup

Joni Juup

CDO, Founder

The HARG Truth: AI's Need for the Human Element

The Human-Augmented Retrieval Generation (HARG) method builds upon the Retrieval Augmented Generation (RAG) model, integrating a crucial human touch to the pipeline.

Valtteri Karesto

Valtteri Karesto

CTO, Founder

Intentface: Human-Centric Computing Through Intent-Driven Interactions

Imagine a future where clicking dropdowns, filling input fields, and browsing through abstract data visualizations are things of the past. That's the future we can build with intentfaces.

Joni Juup

Joni Juup

CDO, Founder