This article was written by Kevin Petrie, VP of Research at BARC, a global research and consulting firm focused on data and analytics.
Picture a chatbot that gives erroneous tax advice, insults a customer, or refuses to issue a justified refund, and you start to appreciate the risks of agentic AI.
This blog, the first in a three-part series, explores why and how organizations must implement new governance controls to address the distinct requirements of AI models and the agents that use them. The second blog will define the must-have characteristics of an agentic AI governance program, and the third blog will recommend criteria to evaluate tools and platforms in this space.
Agentic AI refers to an application (also known as an agent) that uses AI models to make decisions and take actions with little or no human involvement. Agents assess various inputs, then plan and execute sequences of tasks to complete specific objectives. They often delegate tasks to tools, models, or other agents. More sophisticated agents transact with one another, reflect on their work, and iterate to improve outcomes.
Agentic AI represents a compelling new opportunity for digital transformation. Most business, data, and AI leaders now view agents as the ideal vehicle for integrating AI into their business processes. One-third of organizations already have agents in production, according to BARC research, as part of an ambitious push to improve efficiency, enrich user interactions, and gain competitive advantage.
However, agentic AI poses considerable downside if not governed properly, and the downside only worsens as adopters consider more sophisticated and autonomous use cases. Let’s consider the three primary domains of risk — data, models, and agents — then define the new governance controls that models and agents require.
These rising, multiplying risks create an uncertain outlook for data, AI, and business leaders as they weave agents into their business processes. Complexity compounds the problem — each new dataset, model, and agent introduces new interdependencies and points of failure, creating an unwieldy web of things that can go wrong.
While most organizations still struggle with data governance, they at least understand the nature of this longstanding problem. In this blog, we’ll explore the new governance requirements of AI/ML models and agents.
Business, data, and AI stakeholders must address the model risks of toxicity and black-box logic:
1. ToxicityBusiness owners must help data scientists and engineers select appropriate tables, documents, images, and so on that feed into retrieval-augmented generation (RAG) workflows for GenAI. Data scientists then test those models by prompting them with various scenarios and evaluating the resulting outputs. They collaborate with developers to implement rule-based checks (e.g., based on keywords) and evaluator models (e.g., based on sentiment), then alerts or filters that stop impermissible interactions. All these controls need continuous monitoring.
2. Black-Box Logic
Data scientists and engineers can make ML more explainable by implementing interpretable models such as decision trees or linear regressions. They also can use techniques such as SHAP to calculate the impact of features on ML model outputs, or LIME to approximate the relationships of inputs and outputs. Explainability is much more difficult with GenAI models due to the complexity of their underlying neural networks. As a partial solution, data scientists and engineers can satisfy basic customer or auditor concerns by ensuring GenAI workflows cite the sources of their content summaries.
These stakeholders also must address the risks that agents make misguided decisions, take damaging actions, and even develop subversive intentions.
As we learned from Romeo and Juliet, “They stumble that run fast.” Agentic AI is no different. As organizations adopt agentic AI to boost efficiency and innovate, they must also manage new governance risks, from toxic outputs and opaque logic to misguided or subversive agent behavior. Reducing these risks requires cross-functional controls at every level, including safe data selection, explainability techniques, and safeguards like alerts, thresholds, and kill switches. The second blog in our series tackles the next question this raises: How does a modern governance program rise to this challenge?