Breaking News
EXCLUSIVE: “Countering bias in financial AI” – Ash Booth, HSBC; Tanya Kopytina, ACI Worldwide and Gürer Kural, Rabobank in ‘The Fintech Magazine’
Tech billionaire Elon Musk has pledged to create a ‘TruthGPT’ to address the bias that has plagued machine learning and led to perverse outcomes for consumers. Meanwhile, the CEO of OpenAI, which created ChatGPT, has called on US legislators to regulate artificial intelligence.
So, given that AI is now an indispensable technology in financial services, how does the industry ensure it’s working to the benefit of everyone? We put together a global panel of experts to find out. And we started by asking for a definition of AI…That proved to be a tricky question.
Ash Booth, a researcher in machine learning now working for HSBC bank, said that, despite AI being around for six decades, there still isn’t consensus among academics.
“A loose definition that many of us like to use is that it’s a field of computer science that focusses on the development of machines that can think, reason, and perform tasks that humans usually do. To do that, it encompasses different technologies, like machine learning, natural language processing, computer vision, even robotics. Generative AI is part of that subset: it refers to AI systems specifically that create new content, or generate new ideas, based on data they’ve been trained on, and a prompt that’s provided at the time of use.”
While generative AI is stealing all the headlines currently, there have been other major advances in more traditional areas of machine learning that are enabling better decision-making, improved automation, and more personalised customer experiences.One of those is incremental learning, a machine learning method that is well suited for environments in which data changes constantly, making it difficult for other models to keep up – in weather modelling, for example… or fighting fraud.
In 2021, payments solutions company ACI Worldwide received full patent approval for an incremental learning technology that had been developed specifically to give enhanced fraud protection for merchants and financial institutions, and which was embedded into its ACI Fraud Management suite. It’s designed to give the payments industry a fighting chance in a threat landscape where the bad actors’ agile modus operandi, armies of well-remunerated data scientists and superior tech, challenge financial crime teams every day.
ACI’s Tanya Kopytina, took up the story: “Incremental learning allows artificial intelligence to continuously learn and adapt as new data comes into the system, so, in the fraud context, that’s as soon as new fraudulent patterns have been established.
“In the hands of ACI, it’s used to address a number of threats. The first one is in account takeover fraud where criminals gain access to the customer’s account or internet banking details, with the purpose of making unauthorised transactions.
“To mitigate data bias and ‘catastrophic forgetfulness’ we need to develop features more precisely and more accurately, as part of the model development process”
“Another is card fraud. That could include card-not-present fraud, or card-present fraud, using a stolen or counterfeit card, which could be the result of online data breaches, some other point of compromise such as BIN attacks, card testing events, or ATM skimming.
“Last but not least is phishing and social engineering scams. But in all cases, the main benefit is that technology enables us to analyse large volumes of data in real time, and to identify fraudulent patterns as they emerge in real time.”
A BIASED PERSPECTIVE
Beyond fraud detection, there is an ongoing debate, though, as to the inherent risks and advantages in incremental learning models, which highlights wider concerns around bias in artificial intelligence.Ash Booth highlighted some of them:
“When you’re learning on the fly, as new data comes in, you have to be particularly careful about monitoring things like data bias or model bias in real time. That’s not to say it’s not possible, but you have to be very careful about that. Because the algorithm might start to focus too much on recent examples, or you’re just not detecting shifts in bias, that sort of thing.
“Another potential issue is ‘catastrophic forgetting’, which is where your model is focussing too much on that recent data, and can ‘forget’ important trends that it hasn’t seen for a while, particularly when new information and old information start to conflict.”He acknowledged that there are ways to mitigate some of those downsides.
Tanya Kopytina explained ACI’s approach: “In order to enable the incremental machine learning technology for fraud detection, first and foremost we need to have the technological capabilities of a fraud detection platform to receive and process data in real time, but also be able to develop machine learning models within the same platform.
“Secondly, to mitigate data bias and ‘catastrophic forgetfulness’ we need to develop features more precisely and more accurately, as part of the model development process. We need to create dynamic features that would be capable to adapt automatically to the changes in fraudsters’ behaviour. When we put these features into the model, the model will be continuously retrained, in a natural manner, and fraudsters basically will be telling us how the model will develop and evolve over time.
“I think it will depend on a combination of technology, fraud, and data science expertise, to be able to apply AI for fraud detection, and ACI has patented its technology for the implementation of that solution.”
Gürer Kural from Rabobank believed that incremental learning was adding a lot of value to AI systems, but cautioned that, beyond fraud detection, there were inherent problems in applying it to other financial use cases: “When it comes to, let’s say, applying it to credit loans it’s trickier, because you cannot say to a customer, ‘your loan request is declined. But if you try tomorrow, maybe you will get it, because our AI system keeps learning!’.
“That’s something that we need to be careful about, because the financial environment, and the trust in the financial environment, is quite fragile, and we know how easy it is to break things.“So, maybe we can adjust the use cases, and the learning frequency, according to the topic and use case. Because, as Ash mentioned – if you train something on its own data, it will favour its own data and then, once you lose control of the incremental learning, maybe it’s too late, and you will see a model that’s drifted away from its first course – and we don’t want to have model drift or ‘catastrophic forgetfulness.’”
“It’s really about thoughtful governance. Because the statistics and the practices exist to make sure this [bias] doesn’t happen. Where we see it falling down is when the right governance isn’t in place”
One way to safeguard against such an eventuality and make AI accountable is through an obligation on developers to produce specific AI software engineering documentation that validates procedures and will help in identifying unintended bias and explain decisions that a machine is making.
There are well-known examples of groups of people discriminated against by AI, not because the programmer intended it, but because the data on which the AI was based itself obscured a bias. That results in a clear pattern of unfairness emerging. But it’s harder to identify where a random bias has occurred and challenge it. Take the real case of a COO at a UK financial institution who, to his dismay, discovered that his own company’s AI had flagged him as a possible gambler. It had based that on the fact that he frequently used an ATM that happened to be next to a betting shop.
Most ordinary people would never find out why a system had assigned them the label it had. Such cases highlighted the need for true focus on fairness and explainability, said Ash Booth: “Explainability, in the sense of ‘what are the things that are driving my model?’. In this case, it seemed to be really focussing on where cash was withdrawn. And then you should ask, ‘is that fair? Is that a reasonable metric? What spillover might that have? Do I, as a person creating this model, believe that basing our decision on the location of cash machines is fair, reasonable, and ethical? What other implications might that have?’.
“You need to be able to unpack your model in that way, and you should be doing it when you have a human training a model. When these models go live and they’re updated in real time, we need to have the systems in place to monitor for these sorts of biases in real time, too.
“There’s so much you can do to measure for and mitigate bias. The statistics and the practices exist to make sure this bias doesn’t happen. But, ultimately, it’s about thoughtful and sensible governance around training and deploying of AI models, that look for things like these proxy measures. Where we see it falling down is when the right governance isn’t in place.”
That struck a chord with Tanya Kopytina: “I absolutely agree. It’s up to the data scientists to be able to interpret this data – to collect and aggregate the data in a way that helps us to generate some insights from the model development.
“If I come back to fraud, we need to have some fraud expertise, to be able to aggregate the data in a way that we could gather some insights about a typical customer’s trusted behaviour, so we can compare that to the current suspicious or unknown activity. Those insights could be done with the data, though – that’s basically the beauty and the art of machine learning model development.
“The better the data you have, the better your model will be. When it comes to data, it’s quality over quantity”
“Another important point with regards to the fraud detection and incremental learning, is the need for accurate, consistent and complete fraud automapping. If we can provide the feedback about what fraud looks like to the system, our models will be able to generate better insights, as a result.”
Finally, we wanted to know if this was inevitably an area of financial services where the biggest institutions that had access to the largest repositories of data were at an advantage.
Yes – and no – said Ash Booth: “Larger financial institutions have more data available to them, and perhaps that’s giving them an advantage in training larger or more widely representative machine learning models. However, managing massive numbers of diverse datasets is a tricky technological problem.
“If I’m trying to solve a problem, I want to use the least amount of data as possible, and the most structured, easy-to-work-with. So, I think smaller banks can benefit from using machine learning focussed on specific use cases where they have a competitive advantage, or a unique expertise, whether that’s small business lending or niche markets.
“They could take advantage of this early stage of AI by organising their data better than banks that have to contend with a lot of tech debt, and by building their data architecture in an AI-first way.”
Gürer Kural agreed: “There’s a tendency to think that the more data you have, the better your model will be. This is wrong. The better the data you have, the better your model will be.
When it comes to data, it’s quality over quantity.“That means the larger datasets that international banks have, for instance, might be a problem for the developers, because you will lose time and effort untangling it, whereas with the smaller banks, maybe it’s easier.”
For Tanya Kopytina, working in the field of financial crime, there was more advantage in large and small banks sharing data – or at least patterns of data: “ACI has developed a Network Intelligence Module that enables financial institutions to collaborate. So, when a big bank detects fraudulent activity or identifies some point of compromise, they can share these fraud signals with smaller banks, which can use that information in a preventative manner.
“We are not sharing private or sensitive information, it’s more a mathematical formula, but it could really help a smaller organisation to fight fraud more effectively by utilising insights they don’t possess.
“That could also promote fairness and accountability because they can avoid their AI models alerting on an unfairly targeted audience, such as specific demographics. And that actually promotes transparency as well as really disrupting the whole chain of fraudulent activity.
This article was published in The Fintech Magazine Issue 28, Page 68-70
- EXCLUSIVE: “Getting the Message” – Raj Soni and Egill Ingólfsson, Meniga and Michal Panowicz, BCG in ‘The Fintech Magazine’ Read more
- Embedded Finance & Virtual Cards: payabl.’s Breno Oliveira on Building Smarter B2B Payments Read more
- The Tech Capital Expands Global Reach with Launch of LATAM Finance Forum Read more
- How Are Build-vs-Buy Decisions Shaping the Future of Payments? | Volante’s Nihit Ahuja and NatWest’s Lee McNabb Read more
- Why Customers Trust Bueno | Felix Hughes, Director at Bueno Read more