AI threatens to add to the growing wave of fraud but is also helping tackle it.
There were 4.5 million reported incidents of fraud in the UK in 2021/22, up 25% on the year before. It is a growing problem which costs billions of pounds every year.
The COVID pandemic and the cost of living crisis have created ideal conditions for fraudsters to exploit the vulnerability and desperation of many households and businesses. And with the use of AI increasing in general, we will likely see a further increase in new types of fraud and is probably contributing to the increased frequency of fraud we are seeing today.
Already, the ability of AI to absorb personal data, such as emails, photographs, videos and voice recordings to imitate people is proving to be a new and unprecedented challenge.
But there is also an upside. The government, banks and other financial organizations are now fighting back with increasingly sophisticated fraud-detection methods. AI and machine learning models could be a part of the solution to deal with the increasing complexity, sophistication and prevalence of such scams.
The rising gap between prices and people’s incomes appears to have made people more receptive to scams which offer grants, rebates and support payments.
Fraudsters often target individuals by posing as genuine organizations. Examples include pretending to be your bank or posing as the government telling you that you are eligible for a lucrative scheme, in order to steal your identity details and then money.
This follows a dramatic rise in recent years of fraudulent applications to government and regional support packages, mainly implemented in response to the pandemic. Here fraudsters often pose as fake businesses to secure multiple loans or grants.
One of the most outlandish examples of this was a Luton man who posed as a Greggs bakery to swindle three local authorities in England out of almost £200,000 worth of COVID small business grants.
The hurried roll out of such schemes for faster economic impact made it difficult for officials to effectively review applications. The UK government’s Department for Business and Trade now estimates that 11% of such loans, roughly £5 billion, were fraudulent. By March 2022 only £762 million had been recovered.
Fraud detection
Over the past few years, complex mathematical models combining traditional statistical techniques and machine learning analysis have shown promise in the early detection of financial statement fraud. This is when companies typically misrepresent or deceive investors into believing they are more profitable than they really are.
One of the breakthroughs has been the incorporation of both financial and non-financial information into data analysis systems. For example, the risk of fraud decreases if there is better corporate governance and a lower proportion of directors who are also executives.
In a small business context, we can think about this as promoting transparency and making sure that important positions do not have sole authority to make significant decisions.
Such data analytics models can be used to rank applications in terms of potential fraud risk, so that the riskiest applications get additional scrutiny by government officials. We are now starting to see implementations of such systems to tackle universal credit fraud, for example.
Banks, financial services providers and insurers are developing machine-learning models to detect financial fraud too. A Bank of England survey published in October 2022 revealed that 72% of financial services firms are already testing and implementing them.
We are also seeing new collaborations in the industry, with the likes of Deutsche Bank partnering with chip maker Nvidia to embed AI into their fraud detection systems.
Risks of AI systems
However, the advent of new automated AI systems bring with it worries of potential unintended biases within them. In a recent trial of a new AI fraud detection system by the Department of Work and Pensions, campaign groups were worried about potential biases.
A common issue that needs to be overcome with such systems is that they work for the majority of people, but are often biased against minority groups. This means if left unadjusted they are disproportionately more likely to flag applications from ethnic minorities as risky.
But AI systems should not be used as a fully automated process to detect and accuse fraud but rather as a tool to assist assessors. They can help auditors and civil servants, for example, to identify cases where greater scrutiny is required and to reduce processing time.
Laurence Jones, Bangor University and Adrian Gepp, Bangor University
Laurence Jones, Lecturer in Finance, Bangor University and Adrian Gepp, Professor of Data Analytics, Bangor University
This article is republished from The Conversation under a Creative Commons license. Read the original article.