Charles Cipione
Dallas
How can a decision-making algorithm and the environment it operates in be fully investigated? In part, this is determined by which element in the decisioning system is suspected of having an issue. In the first article of this series, we identified several potential components of the decision-making system that could present an issue, namely:
In our experience, investigations of this kind do not proceed in a strict linear fashion, especially when the decisioning system is still in active use. Nonetheless, there is an overarching sequence of steps we would typically follow. There are also situations where the potential issue applies to an active system, creating a need for a parallel process that escalates identified issues in real-time so that tactical fixes can be applied to address these.
Gaining an understanding of the purpose and history of the decision-making algorithm is critical to understanding whether issues have occurred and what the likely impacts of these might be. While this may sound obvious and trivial, getting a clear understanding of this can be challenging, particularly when the algorithm has been in place for a long time, has had many changes to it, or where its function spans multiple parts of the business.
Some of the key questions to ask about the algorithm could include:
An immediate priority in investigations of this nature should be an assessment of the need to preserve the system and its related data. The details of this will always be case-specific, but in our experience the following components should be investigated and considered for preservation:
To fully investigate the system, it is often necessary to have a fully working sandbox environment available for interrogation. Often systems will already have testing environments in place, and these can be cloned to provide a sandbox environment in which the system can be safely interrogated. In addition to the IT challenges of setting up a sandbox environment, a key consideration is to gain assurance that the sandbox version of the algorithm is a reasonably faithful replica of the actual system.
During a high-profile regulatory investigation, we spent a significant period calibrating the algorithm, using historical input and output data, to ensure we had faithfully recreated its behavior, and that we could evidence this to the regulator.
At this point in the investigation, we get to heart of the matter. How you go about investigating an algorithm and the wider system it operates in will be fact-specific, but there are three broad approaches that can be used in conjunction:
Once any issues have been identified it is often important to establish their materiality. It may be possible that you have identified an "edge-case" bug: where a certain set of inputs could cause the algorithm to function incorrectly. But if these inputs never actually occurred there’s less cause for concern than if this was a regular occurrence.
To assess materiality, it is often necessary to undertake a lookback exercise. In this exercise, historical input data is fed through a corrected version of the algorithm. The resulting outputs are then compared to the original outputs that were produced. This approach requires both historical input and output data. When this is not available, it may be possible to create synthetic inputs which can be run through the original and corrected version of the algorithms. While lookbacks can be time-consuming exercises, they have the benefit of being able to quantify the actual impacts of any issues that were identified. When a regulator is involved, this is especially advantageous, as the materiality of the issues can factor into the regulator’s resulting sanctions and/ or censure.
As we referenced in the introduction, as the investigation proceeds, any identified issues should be escalated so that temporary fixes can be implemented. These temporary fixes are often put in place quickly and often include inefficient manual checks and controls.
Once the full list of issues is understood, it is advisable to mature these temporary fixes into more permanent ones that will improve the efficiency of the process and make them more robust in the longer-term. Equally, obtaining an understanding of why the issues arose helps to inform changes in how the system should be managed and governed. It may also hold wider lessons for the organization.
In our final article in this series, we will explore how we might go about investigating AI decision-making algorithms, contrasting this with how we’ve outlined investigating conventional algorithms.
Read the previous article in this series: