WEBVTT
00:00:00.480 --> 00:00:07.520
In this video, we’re gonna look at a neat way that forensic accountants can detect fraud.
00:00:08.160 --> 00:00:15.040
It’s been used to uncover expenses scams, fake research, and falsified accounts.
00:00:15.560 --> 00:00:19.160
And people have got criminal convictions as a result.
00:00:19.680 --> 00:00:23.880
Despite this, lots of people still don’t know about Benford’s law.
00:00:24.440 --> 00:00:34.040
Before we talk about the main topic though, we’re gonna take a bit of a detour back to a time when people didn’t have calculators or computers to carry out complicated calculations for them.
00:00:34.520 --> 00:00:38.160
But they did have a few tricks up their sleeves to make life easier.
00:00:38.520 --> 00:00:58.320
For example, to multiply two large numbers together, rather than carrying out a huge long multiplication calculation, they would look up the logarithms of their numbers in a book of log tables, add them together, then convert the answer back to a regular number using antilog tables.
00:00:59.000 --> 00:01:02.520
And they could also carry out division calculations by subtracting the logarithms.
00:01:03.120 --> 00:01:06.640
It saved a lot of time over long multiplication.
00:01:07.360 --> 00:01:11.920
The story of logarithms and how they work is actually pretty fascinating.
00:01:12.760 --> 00:01:29.640
400 years ago, a man called John Napier spent 20 years creating a huge table which listed the log values of whole numbers up to 10 million, essentially using base one minus 10 to the negative seven to make his calculations easier.
00:01:30.240 --> 00:01:39.000
Then his friend, Henry Briggs, converted it all into logs of base 10, which made it easier for everyone else to use the tables for their calculations.
00:01:39.800 --> 00:01:50.920
Then, back in 1881, an American astronomer called Simon Newcomb was using log tables to do loads of calculations.
00:01:51.520 --> 00:02:02.200
And he noticed that the early pages, containing logarithms of values beginning with one and two, were getting much more worn out through use than the other pages.
00:02:02.880 --> 00:02:08.920
He did a bit of research and published a paper about the probabilities of the first digits taking certain values.
00:02:09.680 --> 00:02:17.760
Later in his career, he suggested that astronomers were almost at the stage where they’d found out everything they could about the night skies.
00:02:18.480 --> 00:02:19.920
And he was very wrong about that.
00:02:20.720 --> 00:02:24.600
But he was right about the uneven distribution of first digits.
00:02:25.120 --> 00:02:49.840
Nearly 60 years later, a physicist called Frank Benford independently noticed the nonuniform distribution of first digits in a whole range of datasets from the populations of towns, the values of physical constants, statistical numbers used in news articles, surface areas of rivers, and more.
00:02:50.440 --> 00:02:54.240
And by now, the world was ready to accept the phenomenon.
00:02:54.600 --> 00:03:01.920
And it became known as Benford’s law, although he called it the law of anomalous numbers.
00:03:02.440 --> 00:03:06.080
And he wasn’t actually the first person to discover it.
00:03:06.440 --> 00:03:12.240
Now this often happens, naming laws after someone other than the first person to discover them.
00:03:12.920 --> 00:03:20.240
And when I did some research for this video, I found out that this phenomenon is known as Stigler’s law of eponymy.
00:03:20.640 --> 00:03:31.040
And ironically, when Stephen Stigler proposed his law of eponymy, he noted that it was something first reported by another person called Robert Merton.
00:03:31.440 --> 00:03:43.120
You’ll be pleased to hear that many people now call Benford’s law “Newcomb-Benford’s law” in an attempt to give the original discoverer due credit.
00:03:43.760 --> 00:03:53.880
But it boils down to the frequency distribution of first, or most significant, digits of certain kinds of data looking roughly like this.
00:03:54.200 --> 00:04:05.080
This means that one is the first digit around 30 percent of the time, while nine is the first digit only about four or five percent of the time.
00:04:05.720 --> 00:04:07.760
And that may initially seem surprising.
00:04:08.200 --> 00:04:11.600
We might expect that all digits are equally likely to occur.
00:04:12.080 --> 00:04:25.760
So numbers beginning with one, two, three, four, five, six, seven, eight, or nine would all be equally prevalent with a probability of a ninth, around 11 percent.
00:04:26.400 --> 00:04:31.920
Now we talked about logarithms to explain how Newcomb first discovered this effect.
00:04:32.520 --> 00:04:41.080
But the formula for working out the expected probability of each possible first digit actually involves logarithms as well.
00:04:41.800 --> 00:04:48.120
The probability that the first digit is 𝑥 is equal to log base 10 of one plus one over 𝑥.
00:04:48.840 --> 00:05:04.920
So the distribution we get with this is quite different to the uniform distribution we’d get if the first digit was equally likely to be one, two, three, four, five, six, seven, eight, or nine.
00:05:05.400 --> 00:05:12.960
But if we think about it for a moment, we can see that this law would only work with certain kinds of distributions of numbers.
00:05:13.720 --> 00:05:20.920
For example, if you were to take adult heights in metres, we would expect many more than 30 percent of them to begin with a one.
00:05:21.520 --> 00:05:27.440
And if you took those same heights in feet, then almost none of them would begin with a one.
00:05:28.080 --> 00:05:41.560
If we restricted ourselves to just looking at the counting numbers between one and 90, then about 12 percent begin with a one, 12 percent with a two, and so on.
00:05:42.240 --> 00:05:45.960
But only two of them, about two percent, begin with a nine.
00:05:46.600 --> 00:05:49.200
The distribution of first digits is uniform apart from nines.
00:05:49.520 --> 00:05:51.640
And Benford’s law doesn’t apply.
00:05:52.160 --> 00:05:58.160
For different ranges of numbers, it seems obvious that fewer numbers will begin with the higher digits.
00:05:58.840 --> 00:06:12.840
The law works best on numbers that span a range of orders of magnitude, that is, powers of 10, so single digit, tens, hundreds, thousands, and so on, and don’t have any artificially applied constraints.
00:06:13.360 --> 00:06:19.120
For example, telephone numbers are all the same length and begin with specific area codes.
00:06:19.440 --> 00:06:26.520
And bank card numbers all have an industry identifier as their first digit, as part of the bank identification number.
00:06:27.160 --> 00:06:36.080
So we wouldn’t expect the law to apply when first digits are constrained by such human-defined rules.
00:06:36.600 --> 00:06:57.680
But when we look at populations of countries, areas of states, house prices, or even a range of statistical numbers chosen at random from articles in a newspaper, they’re likely to be drawn from across a range of orders of magnitude and so will probably follow Benford’s law.
00:06:58.520 --> 00:07:02.840
Using the law can help you check the validity of data you’ve collected.
00:07:03.360 --> 00:07:21.040
For example, if you analyse all the expenses claims for a large organisation and the claims vary in size from very small amounts to many thousands of dollars, then you’d expect something like a Benford’s law frequency of first digits.
00:07:21.760 --> 00:07:30.320
If you find something very different, maybe someone has been putting in false claims using a more uniform distribution of first digit numbers.
00:07:30.960 --> 00:07:44.760
It could be that there’s another reason, like certain rules about what people can claim expenses for or limits on the amount of expenses or even an error in the accounting software.
00:07:45.560 --> 00:07:50.040
But it’s a very useful flag of the unexpected.
00:07:50.800 --> 00:07:54.760
Some mathematical sequences of numbers also follow Benford’s law.
00:07:55.240 --> 00:08:10.080
If you write down all the integer powers of two up to any large number, for example, a hundred or a million or a billion or whatever, you’ll see that the first digits have a frequency distribution similar to that suggested by Benford’s law.
00:08:10.080 --> 00:08:15.160
The same happens with Fibonacci numbers and factorials.
00:08:16.000 --> 00:08:19.600
So this leads us to thinking about why Benford’s law works.
00:08:20.280 --> 00:08:30.760
With the mathematically generated sequences, it can be quite easy to see what’s going on, especially when you’ve got exponential sequences like the powers of two.
00:08:31.360 --> 00:08:37.520
With the integer exponents of two, each term in the sequence is double the previous one.
00:08:38.160 --> 00:08:52.120
And if we plot the values on a logarithmic scale, where equal spaces represent exponentially increasing amounts, as you make your way along the axis, we can see that it’s more likely that the first digit will be one.
00:08:52.880 --> 00:09:04.600
On our logarithmic scale, we can see that the space between one and two is much larger than the space between two and three, which is larger than the space between three and four, and so on.
00:09:05.320 --> 00:09:24.160
So if we plot our powers of two on the scale, two, four, eight, 16, and so on, notice how the steps between subsequent numbers, two to four, four to eight, eight to 16, are equally spaced.
00:09:24.760 --> 00:09:27.000
And that’s because of this logarithmic scale.
00:09:27.720 --> 00:09:32.320
So we’re taking equal-sized steps through this scale.
00:09:32.960 --> 00:09:37.960
And a larger proportion of the regions relate to situations where the first digit is equal to one.
00:09:38.520 --> 00:09:45.560
And smaller and smaller areas represent numbers beginning with two, three, four, and so on.
00:09:46.000 --> 00:09:50.840
As we count them up, more of the powers of two will begin with one.
00:09:51.520 --> 00:09:54.200
More of them will fall into these regions.
00:09:54.880 --> 00:10:05.000
Now let’s think about other naturally occurring statistics, like town populations and why they might follow Benford’s law.
00:10:06.080 --> 00:10:08.600
We count people to work out population.
00:10:09.080 --> 00:10:11.440
So let’s start by thinking about some really small towns.
00:10:12.200 --> 00:10:16.520
Obviously, we need at least one person to constitute a tiny town.
00:10:17.080 --> 00:10:26.160
And if our largest town had a population of one, then 100 percent of towns would have populations with a first digit of one.
00:10:26.960 --> 00:10:33.360
And zero percent would have a first digit of two, three, four, and so on, up to nine.
00:10:34.000 --> 00:10:49.480
If the largest town population was two, then on a random distribution basis, about 50 percent of populations would have a first digit of one and 50 percent would have a first digit of two.
00:10:50.640 --> 00:10:51.720
We’ve got two choices.
00:10:52.360 --> 00:10:57.160
And again, no first digits would be higher than that.
00:10:58.080 --> 00:11:09.840
As the maximum size of the town increases up to nine, the proportion of towns you’d expect to have a first digit of one decreases down to a ninth, about 11 percent.
00:11:10.600 --> 00:11:15.440
Now let’s consider towns with populations of up to 10 people.
00:11:16.160 --> 00:11:22.960
Well, those with one or 10 people have populations with a first digit of one.
00:11:23.720 --> 00:11:28.320
So now two out of the 10 options have a first digit of one.
00:11:29.160 --> 00:11:45.600
Then as we include towns with 11, 12, 13 people, and so on, up to 19, the percentage of potential town populations beginning with one increases up to 58 percent.
00:11:46.440 --> 00:12:01.240
If towns have random populations between one and 19, then there’re 11 ways to get a first digit of one out of 19 different possibilities.
00:12:02.080 --> 00:12:07.160
Then we could think about towns with populations of up to 99 people.
00:12:07.920 --> 00:12:17.200
And the proportion of towns with population first digits of one will reduce down to 11 out of 99.
00:12:17.840 --> 00:12:20.240
That’s just over 11 percent.
00:12:21.080 --> 00:12:48.680
Then as we work our way through possible populations up to 100, then 101, and so on, up to 199, the proportion of possible populations with a first digit of one gradually increases until you see that there are 111 out of 199 ways of having a first digit of one in the numbers up to 199.
00:12:49.440 --> 00:12:52.080
That’s about 56 percent.
00:12:52.880 --> 00:13:01.560
If we plot these proportions on a line graph, we can see that this pattern continues as we increase the possible town size.
00:13:02.160 --> 00:13:24.360
Every time we introduce the next order of magnitude of possible populations, the proportion of possibilities for a first digit of one rapidly increases up to about 50 percent and then decreases slowly back down to 11 percent as we include more of the possible populations up to the next order of magnitude.
00:13:25.080 --> 00:13:42.960
If each of these theoretical maximum populations is equally likely, the expected proportion of town populations beginning with one is some kind of average between about 11 percent and just over 50 percent.
00:13:43.680 --> 00:13:45.720
It turns out to be around 30 percent.
00:13:46.640 --> 00:14:01.960
So if the data we’re looking at is uniformly randomly distributed over the range one to 9999, then we’d expect about 11 percent of the numbers to begin with a one.
00:14:02.720 --> 00:14:16.360
But if the data is uniformly randomly distributed over the range one to 19999, then you’d expect about 56 percent of numbers to begin with one.
00:14:17.040 --> 00:14:33.600
Since both situations are equally likely for a variety of different datasets, then it’s not so surprising that we see an average of around 30 percent of numbers in our newspapers and accounts and general statistics with a first digit of one.
00:14:34.320 --> 00:14:43.640
Benford’s law then isn’t actually a mysterious law of anomalous numbers saying that ones appear much more often than you’d think.
00:14:43.960 --> 00:14:57.640
It’s just a simple observation that, depending on where you start and stop counting, more or fewer of the numbers will begin with one.
00:14:58.320 --> 00:15:05.440
The situations in which Benford’s law breaks down are those where we approach either extreme.
00:15:05.960 --> 00:15:18.160
If the maximum possible value for our data is right on the boundary of an order of magnitude, then you shouldn’t be at all surprised to see only 11 percent of your figures beginning with one.
00:15:18.760 --> 00:15:29.640
But if the maximum possible value is about 10 percent higher than that, then over 50 percent of your numbers could begin with one.
00:15:29.640 --> 00:15:39.400
So Benford’s law is an observation that more numbers representing statistical observations tend to begin with lower digits than higher digits.
00:15:40.040 --> 00:15:45.680
And this can help us to spot when someone is trying to cook the books or falsify their data.
00:15:46.280 --> 00:15:55.360
When used wisely, it can help us validate and verify our data and prove a really useful scientific tool.