@mjambon Yea its a bit harder to understand because its a bit of applying it in the opposite way you normally think of it.
So regression to the mean is explained one way that i think is not particularly counter intuitive to how it is applies but let me start with the basics.
Formally, regression to the mean is all about how if you take lots of samples of things with all sorts of bizzare distributions, in the end they will eventually average out to a normal distribution (thus explaining why normal distributions tend to be the default and crop up everyone)...
In practice though the fallacy aspect arrises when you pay attention to addressing outliers, and on resampling they appear to have been "fixed", when in reality they only cropped up as outliers in the first place due to random chance and nothing as changed.
A very typical example given is if you look at a city and pick the top 5 intersections where accidents took place last year and put additional safety measures in place the next year you will notice that those intersections have reduced the number of accidents significantly. You assume this is due to your safety measures when in fact that would have happened regardless since they were only outliers by random chance and they simple "regressed to the mean"...
So how does that apply here. Well like i said its a bit of what i just said but kinda in reverse. You are assuming your sampling is average, when in fact you are samping outliers. So while the reality tends to regress towards the mean (went home after their normal average length bike ride) those that remain are the outliers but you dont recognize them as outliers. So its the same principle of regression to the mean just, the inverse of it.
Make sense now?