This working memo aims to understand what is meant by "AI agent", how it is defined in the wild and how to classify and categorise them?
Defining AI agent
When looking up for definitions of an AI agent, one finds similar, yet slightly different descriptions. An article from IBM describes AI agent as "a system or program that is capable of autonomously performing tasks on behalf of a user or another system by designing its workflow and utilizing available tools [1]. They also make a distinction between agentic and non-agentic: Agentic AI chatbots learn to adapt to user expectations over time. Non-agentic AI chatbots are ones without available tools, memory or reasoning.
AWS says (AI) agent is a software program that can interact with its environment, collect data, and use the data to perform self-determined tasks to meet predetermined goals. Humans set goals, but an AI agent independently chooses the best actions it needs to perform to achieve those goals [2].
Google puts it as follows: AI agents are software systems that use AI to pursue goals and complete tasks on behalf of users. They show reasoning, planning, and memory and have a level of autonomy to make decisions, learn, and adapt [4].
Anthropic goes a bit deeper and practical in their article and associated Youtube video https://youtu.be/LP5OCa20Zpg. They make an architectural distinction between workflows and agents: "workflows are systems where LLMs and tools are orchestrated through predefined code paths. Agents are systems where LLMs dynamically direct their own process and tool usage, maintaining control over how they accomplish tasks" [8].
Consolidating the description
Consolidated characteristics of an AI agent:
A computer system (program or programs)
Works for user or system
Works to meet predetermined human set goals
Autonomously determines best action (task) for itself and path forward and path forward
Autonomously utilises available tools (send commands or fetch new data)
Autonomously learn and adapt
What I think is missing from the descriptions:
How is the quality of outputs assessed and managed?
What is the human involvement in assessing the outcomes?
When is the work complete? What is the life-cycle of an agent?
Which of the AI agent characteristics requires llm to function?
Does user need to know about the existence of AI agents or can they operate behind the scenes?
It could be useful to make a clear distinction between LLM based agents and other type of AI agents.
My own running definition of a LLM based AI agent:
An AI agent in the context of LLM is a program that achieves predetermined goals by defining its own path from prompt to prompt, incorporating additional logic. It can be invoked immediately, on a schedule, or from an event.
Marketed benefits
Improved productivity: autonomous systems achieves business goals without human intervention. By eliminates repetetion adds value to the organisation [2].
Reduced costs: reduce costs by removing inefficiencies, human errors and manual processes through automation [1][2].
Informed decision-making: advanced use of data to have better business decisions and predictions [2].
Improved customer experience: personalized experiences through AI agents improve customer engagement, conversion and loyalty [2].
Risks and limitations
Managing data privacy requirements [2]
Complex agents might require a lot of data to work [2]
Ethical challenges in unfair, biased or inaccurate outcomes [2]
Multiple agents interacting together might become hard to debug [1]
Vendor-locks and running costs
Infinite feedback loops [1]
Complex agent setups might become computationally complex and expensive [1][2]
Adding knowledge / tools
An agent often requires information about the world in order to process the task at hand. This is possible my providing the agent a set of tools that can include external data, web search, APIs or other agents (and their tools). Once the needed information is retrieved the agent's "knowledge" can be updated.[1]
Note: (at the time of writing) not all vendors and APIs supports tools calling.
Types
Simple reflex agents [1][2]
No memory
No interaction to other agents
Rule/condition-based
Not capable to handle unexpected events
Example: suggest password change based on keywords in user conversation
Example: A thermostat that turns on the heating system at a set time every night.
Model-based reflex agents [1][2]
Use current perception and memory to maintain an internal model of the world.
Model is updated once new information is received
Stores information in memory and can operate in environments that are partially observable and changing.
Limited by own set of rules.
Example: A robot vacuum cleaner.
Similar to simple reflex agents, but more advanced decision-making mechanisms.
Can evaluates probable outcomes and consequences before deciding.
Goal-based agents [1][2]
Has internal model of the world and also a goal or set of goals.
Searches for action sequences that reach their goal and plan these actions before acting on them.
Example: A navigation system that recommends the fastest route to your destination.
Has reasoning capabilities.
Evaluating environment data and compares different approaches to help it achieve the desired outcome.
Suitable for performing complex tasks that requires reasoning.
Utility-based agents [1][2]
Select the sequence of actions that reach the goal and also maximize utility or reward.
Utility is calculated using a utility function (pre-fixed criteria).
The criteria includes factors e.g. progression toward the goal, time requirements, or computational complexity.
Agent selects the actions that maximize the expected utility.
Example: A navigation system focusing fuel efficiency and minimizes the time spent in traffic and the cost of tolls.
Learning agents [1][2]
Similar to other agent types but can also learn.
New experiences are added to their initial knowledge base, which occurs autonomously.
Can operate in unfamiliar environments.
Learning agents may be utility or goal-based in their reasoning
A learning agent continuously learns from previous experiences to improve its results.
Example: Personalized recommendations on e-commerce sites.
Hierarchical agents [2]
Organized group of intelligent agents arranged in tiers.
The higher-level agents deconstruct complex tasks into smaller ones and assign them to lower-level agents.
Each agent runs independently and submits a progress report to its supervising agent.
The higher-level agent collects the results and coordinates subordinate agents to ensure they collectively achieve goals.
Architectures
No one standard architecture exists for building AI agents [1]. Some paradigms:
ReAct (Reasoning and Action) [1]
Agents "think" and plan after each action taken and with each tool response to decide which tool to use next.
Think-Act-Observe loops are used to solve problems step by step and iteratively improve upon responses.
"Chain-of-Thought"
ReWOO (Reasoning WithOut Observation) [1]
Eliminates the dependence on tool outputs for action planning: agents plan upfront.
Redundant tool usage is avoided by anticipating which tools to use upon receiving the initial prompt from the user.
User can confirm the plan before it is executed.
ReWOO workflow is made up of three parts:
The agent anticipates its next steps given a user's prompt.
Collect outputs produced by calling tools.
Combine initial plan with the tool outputs to formulate a response.
Best practices
Activity logs that saves the discussions threads [1]
Emergency brakes for interruptibility and human supervision [1]
Unique agent identifiers to manage who did what [1]
Anthropic states that "the most successful implementations weren't using complex frameworks or specialized libraries. Instead, they were building with simple, composable patterns" [8]. Their recommendation is to use simplest solution possible and iterate forward from there.