Great video on Benford’s Law here. Matt goes into a good amount of detail outlining why this occurs, why it doesn’t always apply, and what it means if data does/doesn’t follow the Law.
Here is an alternative Piped link(s): https://piped.video/etx0k1nLn78
Piped is a privacy-respecting open-source alternative frontend to YouTube.
I’m open-source, check me out at GitHub.
Removed by mod
What the heck is real life number
An actual measured data point, as opposed to a randomly generated number. Also this principle applies specifically to the first digit. Overall the title is a complete mess.
Basically, when you gather a bunch of data points about real world quantitative phenomena (e.g. town population, lake surface area, etc), you find this distribution curve of leading digits where 1 is something like 30% most frequent, gradually decreasing down to 9 being least frequent.
This is called Benford’s Law, it’s basically an emergent property about how orders of magnitude work. It’s useful because you can use it to detect fake data, since if your data faker doesn’t know about it they’ll generate fake data that looks random but doesn’t follow this distribution.
something that isn’t an imaginary life number
This is used to catch tax fraud. People who forge reciepts tend to use random numbers, so they stand out as outliers, and they get caught that way.
Removed by mod
This is a bit weird. I was just listening to Infinity 2 today (great book. Totally recommend), and there’s a section where the characters use Benford’s Law to prove reality. I then had to look it up myself.
Just a super weird coincidence…unless Lemmy is listening to me…
This is called the Baader–Meinhof phenomenon, or frequency illusion.
Does anybody know if this is a feature of a decimal system?
I think it’s a feature of all positional notation systems.
The distribution shown in this post is for base 10, but Benford’s Law includes distributions for other bases too. The wiki article linked in another comment goes into detail on that too.
The percentages change. At the lower end, in binary every number that isn’t 0 itself starts with a 1.
This fact is actually used to save one bit in the format that computers usually use to store floating point (fractional instead of integer) numbers.
If you were in Base 12 or something it would still lean towards 1 but the percentage would be a little different.
but… why?
https://en.m.wikipedia.org/wiki/Benford's_law
Look at the logarithmic scale. This law has to do with number sets in the wild, so apparently the scaling is flat over the set of data they examined. If you look at the distribution of the number sets over the logarithmic scale, they are evenly distributed. If you looked at the same numbers on a linear scale, they would become more and more sparse as they grow in size.
Cool! Now imagine I’ve got severe brain damage… can you explain that again?
The further left you are in a number, the more likely it is that the digit will be small
What a great explanation! Thanks for dumbing it down.
Removed by mod
In real life distributions you are always going to have situations where you fill up the bigger digits last, so it becomes less likely they show up. The best example of this is the population of cities. For cities between 100k and 999k you’ll have a larger number of cities with 100k-300k because cities of those sizes are smaller and more common.
Benford’s law is about the leading digit, so it doesn’t matter if the numbers are rounded or not.
No problem!
So if you have a small amount of something, you’ll have maybe 2, maybe 3, or 4, or 5, or 6, 7, 8, 9, 10, 11, 12, 13, 14 or so. If you have a medium amount of something, the numbers might be 20, or 30 ish, or 40 ish, 50s, 60s, 70s, 80s, 90s, 100ish, 110ish, 120 or so, around 130. Larger amounts of stuff end up being 200ish, 300, 400, 500, 600, 700, 800, 900, 1000ish, around 1100, 1200 something, 1300
All the numbers I’ve mentioned are about evenly spaced on this logarithmic scale. You can see that a bunch of them start with 1 just because of how big we think they are! It turns out there is a math reason for this, instead of just being about the weird way humans think.
deleted by creator
So if I rolled a 10 sided dice 1000 times 30% of those rolls would be a 1?
No
Thanks. Now I understand
From what I understand it works like this.
Let’s say you have a series of numbers that represent real life data. In general the first number of all of these numbers will be a 1, 30% of the time.
Such as “1000 rolls”
It applies to situations with more than one order of magnitude being counted, such as d20 rolls, 55% of which will start with a 1.
The “1” of the “1000” is the real life number. He didn’t pick “785” or “462”.
I don’t know why you’re being downvoted.
Thanks, that makes sense. I must be missing a link or article on my client otherwise I would’ve read it lol
It works on things that operate on a logarithmic scale. It’s odd how many real-world things fit that mold that don’t intuitively seem like they would.
Another factor promoting it in real-world data sets is that they often have restricted ranges that favor lower numbers. Days of the month, for example, only go from 1 to 31. There’s only one way for the leading digit to be 4, but there are eleven ways for the leading digit to be 1.
Another type of data includes values of varying ranges, which also favors lower leading numbers. Street numbers start at 1 and go up, ending at some point within a fairly large range in the real world. All of these ranges will have their fair share of leading 1s. They will NOT all have a fair share of leading 2s (what if it ended before 20?), and as you go up it gets progressively less likely. So if you took all street addresses, you’d expect to see more leading 1s than 9s.
Your theoretical dice roll is not such a case. You would expect a uniform distribution of leading numbers. This would hold true with a 99-sided die as well.
While that’s true with a 10-sided die 20% of your rolls will start with a one and all other digits only have a 10% chance.
Oh, yes. Thanks for the correction!
No it is a property of real life thing. It come from the fact that most thing in real world, dont go over 30 or 300 so often. Like number of houses in a street.
Digits, not numbers.