HomeArtificial IntelligenceGetting Started with AI Agents (Part 2): Autonomy, Safeguards, and Pitfalls

Getting Started with AI Agents (Part 2): Autonomy, Safeguards, and Pitfalls

In ours first installmentWe have outlined key strategies for using AI agents to enhance business efficiency. I explained how agents, unlike standalone AI models, iteratively refine tasks using context and tools to enhance outcomes akin to code generation. I also discussed how multi-agent systems promote cross-departmental communication, create a consistent user experience, and increase productivity, resiliency, and faster upgrades.

Success in constructing these systems is dependent upon assigning roles and workflows, in addition to establishing safeguards akin to human oversight and error controls to make sure protected operations. Let’s dive into these crucial elements.

Safeguards and autonomy

Since agents imply autonomy, in a multi-agent system, various protections have to be built into an agent to scale back errors, waste, legal risks, or damages when agents operate autonomously. Applying all of those protections to all agents could also be excessive and resource-challenging, but I strongly recommend considering each agent within the system and consciously deciding which of those protections they would want. An agent shouldn’t be allowed to act autonomously if any of those conditions are met.

Explicitly defined human intervention conditions

Triggering any predefined rule determines the conditions under which a human must validate an agent's behavior. These rules must be defined on a case-by-case basis and will be declared within the agent's system prompt – or, in additional critical use cases, enforced using deterministic code outside the agent. One such rule for a buyer can be: “Every purchase should first be checked and confirmed by a human.” Call your function “check_with_human” and only proceed if it returns a worth.”

protection agents

A protection agent will be paired with an agent tasked with checking for dangerous, unethical, or non-compliant behavior. The agent will be forced to at all times check all or certain elements of his behavior against a protection agent and never to proceed unless the protection agent gives the green light.

uncertainty

Our lab recently published one Paper on a method that may provide a measure of the uncertainty of what generates a big language model (LLM). Given LLMs' propensity for confabulation (commonly often known as hallucinations), favoring a specific output could make an agent significantly more reliable. There are costs here too. To assess uncertainty, we’d like to generate multiple outputs for a similar request in order that we are able to rank them based on certainty and choose the behavior that has the least uncertainty. This can decelerate the system and increase costs, so this must be considered for more critical agents throughout the system.

Unlock button

There could also be times when we’d like to stop all autonomous agent-based processes. This could also be because we’d like consistency or we’ve got identified a behavior within the system that should be stopped while we work out what’s flawed and how one can fix it. For more critical workflows and processes, it is crucial that this separation doesn’t lead to all processes stopping or becoming entirely manual. Therefore, it is strongly recommended to supply a deterministic fallback operation mode.

Work orders generated by the agent

Not all agents inside an agent network should be fully integrated with apps and APIs. This can take some time and requires just a few iterations to get right. My advice is so as to add a generic placeholder tool to agents (normally leaf nodes within the network) that simply outputs a report or work order containing suggested actions to be performed manually on behalf of the agent. This is a fantastic strategy to bootstrap and operationalize your agent network in an agile way.

Test

With LLM-based agents we gain robustness on the expense of consistency. Due to the opacity of LLMs, we’re also coping with black box nodes in a workflow. This signifies that we’d like a distinct testing regime for agent-based systems than for traditional software. The excellent news, nonetheless, is that we’re used to testing such systems because we’ve got been running human-driven organizations and workflows because the starting of industrialization.

While the examples I showed above have a single entry point, all agents in a multi-agent system have an LLM as a brain and might subsequently act as an entry point to the system. We should use divide and conquer and first test subsets of the system ranging from different nodes throughout the hierarchy.

We may use generative AI to develop test cases that we are able to run on the network to research its behavior and make it reveal its weaknesses.

Finally, I’m a giant proponent of sandboxing. Such systems should initially be introduced on a smaller scale in a controlled and secure environment before being phased in to switch existing workflows.

Fine-tune

A standard misconception about genetic AI is that the more you employ it, the higher it gets. This is clearly flawed. LLMs are pre-trained. Nevertheless, fine-tuning them can influence their behavior in various ways. Once a multi-agent system has been developed, we are able to select to enhance its behavior by taking the logs from each agent and labeling our preferences to create a fine-tuning corpus.

Pitfalls

Systems with multiple agents can go right into a tailspin, leading to a question occasionally never completing and agents continually communicating with one another. This requires some kind of timeout mechanism. For example, we are able to check the communication history for a similar query. If it gets too big or we notice repetitive behavior, we are able to stop the flow and begin again.

Another problem that may arise is a phenomenon I prefer to call overload: expecting an excessive amount of from a single agent. The current state-of-the-art for LLMs doesn’t allow us to provide agents long and detailed instructions and expect them to follow them in any respect times. Did I also mention that these systems will be inconsistent?

A workaround for these situations is what I call granularization: splitting agents into multiple connected agents. This reduces the load on each agent and makes the agents' behavior more consistent and fewer prone to go right into a tailspin. (An interesting area of ​​research our lab is conducting is automating the granularization process.)

Another common problem within the design of multi-agent systems is the tendency to define a coordinator agent that invokes different agents to perform a task. This creates a single point of failure, which might result in a reasonably complex structure of roles and responsibilities. In these cases, I suggest pondering of the workflow as a pipeline, with one agent doing a little work after which passing it on to the following.

Multi-agent systems also are likely to pass context to other agents. This can overload and confuse the opposite agents and is commonly unnecessary. I suggest allowing agents to keep up their very own context and reset the context after we know we’re coping with a brand new request (much like how sessions work for web sites).

Finally, it is crucial to notice that the bar for the capabilities of the LLM, which serves because the brains of the agents, is comparatively high. Smaller LLMs may require quite a lot of timely engineering or fine-tuning to fulfill requirements. The excellent news is that there are already several business and open source agents, albeit relatively large ones, that pass the bar.

This signifies that cost and speed have to be a very important consideration when constructing a multi-agent system at scale. It also needs to be assumed that while these systems are faster than humans, they are usually not as fast because the software systems we’re used to.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Must Read