Identifying the normal (or even the abnormal) in mathematics can pose serious difficulties.
In 1909, mathematician Émile Borel (1871–1956) introduced the concept of normality as one way to characterize the resemblance between the digits of a mathematical constant such as pi (the ratio of a circle’s circumference to its diameter) and a sequence of random numbers.
If a number is normal, digit sequences of the same length occur with the same frequency. A constant would be considered normal to base 10 if any single digit in its decimal expansion appears one-tenth of the time, any two-digit combination one-hundredth of the time, any three-digit combination one-thousandth of the time, and so on.