Firms worry that lack of ‘explainability’ will be regulatory roadblock for AI

Industry experts share their concerns about advanced AI’s ‘black-box’ nature and how that may attract fragmented regulatory scrutiny.

Those responsible for leading AI initiatives at financial firms fear that the opaque nature of generative AI models and the large language models (LLMs) on which they are trained may significantly delay its uptake for certain tasks as regulators seek greater clarity into its use and operation.

At last week’s North American Financial Information Summit, hosted by WatersTechnology, speakers on different panels repeatedly cited the challenge of explainability—i.e., the ability to explain how and why an AI model generates a particular result or arrives at a decision.

While only a few years ago firms were holding back on AI development, citing a lack of regulatory guidance, the current push to move quickly to take advantage of GenAI is resulting in a rushed and fragmented response from regulators, leading to fears that firms will be asked to comply with multiple and inconsistent rules from different jurisdictions, and that these rules will demand greater transparency into AI models and LLMs, which they may not be able to provide.

This amounts to knowing what technology platforms your firm is using, knowing how those platforms work, and being able to demonstrate that knowledge and understanding to customers—because customers also want to know how the sausage is made—and regulators, said Elena Treshcheva, program manger at Exactpro systems.

Tarun Sood, chief data officer at American Century Investments, said that while traditional AI and GenAI share certain key similarities, such as the importance of data quality and good data management, where GenAI differs is this explainability factor.

Other speakers also argued that GenAI can’t be treated in the same way as traditional AI: it’s smarter and more complex. But with that comes additional challenges, said Kristof Horompoly, former executive director and head of responsible AI at JPMorgan Chase.

“This is a shift in the industry mindset. Until now, we’ve had high expectations of AI and machine learning models to be explainable in some way … and to have insight into the underlying data points. But when you move into the GenAI space, you have to relax those expectations, and so for anything that’s regulated, you can’t use it for those tasks now.” He recently left the bank and earlier this month joined ValidMind, a Palo Alto, California-based provider of technology that performs automated documentation and validation of AI models for risk and regulatory purposes.

“There could be some use cases that are off limits to GenAI forever—which I find unlikely—or we figure out transparency and explainability,” said Horompoly, who serves as the company’s vice president of AI risk management. “Or we use the models’ performance over time to demonstrate they’re safe to use without fully understanding them.”

In theory, a firm could record and present the code used to build the model, the data sources used as inputs, the questions used to train it, and the results it generates, as well as how those results change over time as the model learns. But, warned Horompoly, “If you are using [an AI tool built by] a vendor, you have to accept that you’ll have no insight into what data was used to train a model.”

Weighing the trade-offs

Beyond potential regulatory interest, this lack of insight or explainability also poses challenges to firms managing these projects and trying to control and contain them in-house, requiring them to move more slowly than they would ideally like in the interests of reliability, accuracy and security—not to mention the risk of a model leaking confidential information, such as customer data, which would lead to reputational damage and also attract the attention of regulators.

“Traders want to build fast, because in financial services, speed is alpha,” Sood said. “We don’t want to halt innovation, but we have to safeguard our data.”

Vaibhav Bang, vice president of compliance advisory at Societe Generale, agreed, urging a cautious and controlled approach to rolling out AI models. “The saying, ‘Go fast and build things.’ shouldn’t be applied in this case. The expanded use of AI has to be done in a very slow and ethical fashion.”

Traders want to build fast, because in financial services, speed is alpha. We don’t want to halt innovation, but we have to safeguard our data
Tarun Sood, American Century Investments

But establishing guidelines and rules to keep financial firms on the straight and narrow and consistent with each other would require—at the very least—cooperation between regulators, if not the establishment of a global rulemaking body for the use of AI in finance. Rather than dealing with fragmented regulatory regimes, Horompoly advocated for a more global approach and standards, but said that any governing body to meaningfully address AI must do so without barriers between jurisdictions.

“Any governing body will need to be international, and will need to be powerful,” he said.

While much of the AI regulation enacted so far has focused on consumer privacy and protection, as well as IP protection, and combatting deepfakes, rather than anything specifically emerging from authorities that regulate capital markets, a large part of firms’ concerns in the meantime is who will be using AI within their organizations, and what level of understanding they have about it.

“We’re seeing a much broader audience using these [GenAI] implementations,” Horompoly said. “Traditional models were very quantitative and were used by subject matter experts to build quant models. But we’re seeing GenAI users across a whole company with different levels of understanding. So, what’s becoming more important is education. There might need to be training before people start using it; the risks and responsibilities need to be made clear; and people need to not take outputs for granted.”

Barriers to explainability

It’s the sometimes inconsistent and inaccurate results of AI—called hallucinations—that require people to still check results without relying blindly on them, or risk suffering monetary loss, reputational damage, or penalties if they use incorrect data generated by AI. These hallucinations are also a barrier to explainability, especially if the industry adopts the performance-monitoring option for demonstrating reliability.

How do we get rid of hallucinations? We probably don’t. Models always make errors. It’s about how you deal with hallucinations
Kristof Horompoly, ValidMind

“A lot of the ongoing monitoring will have to come from hard-coded guardrails or filters—either from a vendor or internally—and there needs to be a feedback channel for users to report issues,” Horompoly says. And while there are ethical questions surrounding the use of filters for protection—such as, who decides what type or level of filter to impose—the fact remains that introducing any filter will slightly degrade a model, so this approach will always involve a trade-off, he warns.

“How do we get rid of hallucinations? We probably don’t. Models always make errors. It’s about how you deal with hallucinations,” Horompoly says. One way could be to run models more than once, since it’s unlikely a model will make the same mistake twice. Another is to have human subject matter experts in the loop to spot hallucinations—though Horompoly warns that these approaches will result in significant increases to the cost of using AI.

He and other speakers all emphasized the importance of keeping people—specifically, staff with expert knowledge of the area being addressed—in the loop who have the experience to identify and monitor the AI’s successes and failures.

“The best risk management is an educated user,” said Iro Tasitsiomi, head of investments data science at T. Rowe Price.

Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.

To access these options, along with all other subscription benefits, please contact info@waterstechnology.com or view our subscription options here: http://subscriptions.waterstechnology.com/subscribe

You are currently unable to copy this content. Please contact info@waterstechnology.com to find out more.

Most read articles loading...

You need to sign in to use this feature. If you don’t have a WatersTechnology account, please register for a trial.

Sign in
You are currently on corporate access.

To use this feature you will need an individual account. If you have one already please sign in.

Sign in.

Alternatively you can request an individual account here