# Normal Distribution

The **normal distribution** is a probability distribution that (roughly) describes many common datasets in the real world. It is the most common type of distribution, and it arises naturally in statistics through random sampling techniques.

The normal distribution, also known as the **Gaussian distribution**, originally arose in the context of estimating human and mechanical error in the use of scientific equipment, like telescopes. Nowadays, it is more common to show up as a model for the "lifespan" of a product, like a lightbulb, or the outcome of standardized tests, like IQ. Biological measurements, like height or weight, are often estimated with normal distributions (though log-normal may be more accurate), and stock values were originally estimated in the same way (though they tend to have fatter tails).

The plot of a normal distribution is symmetric about the mean and has skinny tails, which are less formal ways of saying its skewness and kurtosis are each \(0\). A given normal distribution is then characterized entirely by the values of its mean and standard deviation. The median and mode of a normal distribution are equal to its mean.

## Normality

The central limit theorem states that the probability distribution of the average of \(n\) independent, identically distributed random variables with finite variance converges to the normal distribution as \(n\) approaches infinity. No matter what the distribution of the random variables are (even if it is very non-normal), the distribution of sample means is normal! In empirical cases, \(n = 30\) suffices for a sufficiently "normal-like" distribution. (This number was determined by computers running way too many Monte Carlo simulations.)

So, even if a distribution is not normal, the distribution of its sample means (for sample size \(>\) 30) *is* normal! With this in mind, it becomes more reasonable that normal (and normal-like) distributions appear so much in nature or the real world.

Many common examples of measurements in statistics can, from another perspective, be thought of as the combination of a multitude of independent factors. For instance, a person's height is the result of a complicated jumble of genetic, environmental, and personal factors. The accuracy of measurements from a telescope may be dependent upon the eyesight or steady-handedness of the astronomer, the effectiveness of the mechanisms that constructed the telescope, the presence of microscopic dust on the lens, and atmospheric interference from weather, gravity, and background noise. While these examples do not meet the explicit criteria for the **central limit theorem**, they hopefully at least show why it might be intuitive that the normal distribution appears so often: it comes from a synthesis of other distributions.

It is worth noting that while many standard measurements in social sciences and elsewhere are often assumed to be normal, they may not be in reality. However, the distribution of averages from simple random samples is indeed normal.

Suppose all \(2000\) students at a school each wrote on a slip of paper their favorite number from \(1\) to \(100\). Then, all the numbers were collected and put in a box. The math teacher then chose \(30\) slips at random, wrote down the average of the numbers written on them, and put the slips back in the box. The teacher then does this many, many more times. No matter what the probability distribution was for the individual students, the probability distribution for the numbers the math teacher writes down is normal! If those numbers were plotted, the teacher would get the typical bell curve.

The takeaway is that many naturally occurring processes are susceptible to statistical techniques that work for normal distributions. In particular, even very non-normal distributions (with finite variance) have a sampling distribution of means that is normal.

However, the assumption of normal distribution fails to hold in many cases. Here are some good rules of thumb for evaluating whether or not it is reasonable to assume normality:

Holds | Fails |

Amalgamation of similar distributions | Dominated by one (or few) particular distribution |

Contributing factors are independent | Dependencies among contributing factors |

Sample selection is uniformly random | Sample selection is correlated to previous selection |

There are statistical tests of the normality of a probability distribution, like the Kolmogorov-Smirnov test, the Shapiro-Wilk test, and Pearon's chi-squared test. These attempt to disprove a null hypothesis that the distribution can be modeled by a normal distribution by finding sufficient evidence that the distribution departs from a normal distribution. The quantile-quantile plot provides a popular and graphical technique for determining whether two datasets have the same distribution. This requires independent sampling from a known normal distribution.

Student scores on history quizzes are likely to be non-normal, since their performance is dominated by whether or not they read the material before class. The distribution is likely to be left-skewed.

The 2008 financial crisis was arguably caused by long-term adherence to the assumption that stock prices are normal when, in fact, there is often a herd mentality contributing to swift rises/falls in price. Dependencies among contributing factors lead to distributions with fatter tails than the normal distribution.

In many empirical studies (for instance, surveying a group of people), it is important that respondents do not hear other participant's answers before giving their own. This is because people may change their answer based upon what they hear, and the previous person may have affected the next person's response, skewing results even if the people were selected at random.

Overall, statisticians are well-disposed to talk about whether a distribution is normal or not.

## Properties

The normal distribution has two important properties that make it special as a probability distribution.

The average of \(n\) normal distributions is normal, regardless of \(n\).

There exist other distributions that have this property, and they are called *stable distributions*. However, the normal distribution is the only stable distribution that is symmetric and has finite variance. Such sums are known as multivariate normal distributions.

Given a simple random sample from a random variable with a normal distribution, the sample mean and sample variance are independent.

This property is unique (among all probability distributions) to the normal distribution. It emphasizes the overall symmetry and "balance" of the bell curve.

Histograms show how samples of a normally distributed random variable approach a bell curve as the sample size increases. The following graphs are of samplings of a random variable with normal distribution of mean \(0\) and standard deviation \(1\).

Note how the graphs become more and more symmetric as \(n\) increases. The proportion of numbers in a certain region also begin to have fixed ratios. For instance, \(68\%\) of the numbers in the last graph appear between \(-1\) and \(1\). In fact, all normal distributions have these same ratios, and tables of \(z\)-scores are used to determine the exact proportions.

A new product was released and a survey asked customers to give the product a score between 1 and 100. At first, when the number of subjects \((n)\) was still relatively low, the company couldn't pull much information from the surveys. For example, after four people had taken the survey, one person rated it a 92, one rated it a 72, one rated it a 63, and the last one rated it a 34. However, as more customers took the survey, the company was able to create a histogram showing the results. Once 5,000 surveys had been taken, the company found that the average person rated the product a 67 out of 100, and the rest of the scores were normally distributed in a bell-curve out from there (with a standard deviation of 9). Based on this, the company decided that its product was not meeting customers' desires.

## z-scores

The \(z\)-score of a value from a random variable is the number of standard deviations away from the mean the value is. If \(\sigma\) is the standard deviation of the distribution, \(\mu\) is the mean of the distribution, and \(x\) is the value, then \[z = \frac{x - \mu}{\sigma}.\] This value is used in many tests in statistics, most commonly the \(z\)-test. By calculating the area under the bell curve, a \(z\)-score provides the probability of a random variable with this distribution having a value less than the \(z\)-score.

A \(z\)-score table usually takes the following form, where the column determines the hundredths digit of the \(z\)-score and the row determines the tenths and unit digit.

\(z\) | .00 | .01 | .02 | .03 | .04 | .05 | .06 | .07 | .08 | .09 |

–3.4 | .0003 | .0003 | .0003 | .0003 | .0003 | .0003 | .0003 | .0003 | .0003 | .0002 |

–3.3 | .0005 | .0005 | .0005 | .0004 | .0004 | .0004 | .0004 | .0004 | .0004 | .0003 |

–3.2 | .0007 | .0007 | .0006 | .0006 | .0006 | .0006 | .0006 | .0005 | .0005 | .0005 |

–3.1 | .0010 | .0009 | .0009 | .0009 | .0008 | .0008 | .0008 | .0008 | .0007 | .0007 |

–3.0 | .0013 | .0013 | .0013 | .0012 | .0012 | .0011 | .0011 | .0011 | .0010 | .0010 |

–2.9 | .0019 | .0018 | .0018 | .0017 | .0016 | .0016 | .0015 | .0015 | .0014 | .0014 |

–2.8 | .0026 | .0025 | .0024 | .0023 | .0023 | .0022 | .0021 | .0021 | .0020 | .0019 |

–2.7 | .0035 | .0034 | .0033 | .0032 | .0031 | .0030 | .0029 | .0028 | .0027 | .0026 |

–2.6 | .0047 | .0045 | .0044 | .0043 | .0041 | .0040 | .0039 | .0038 | .0037 | .0036 |

–2.5 | .0062 | .0060 | .0059 | .0057 | .0055 | .0054 | .0052 | .0051 | .0049 | .0048 |

–2.4 | .0082 | .0080 | .0078 | .0075 | .0073 | .0071 | .0069 | .0068 | .0066 | .0064 |

–2.3 | .0107 | .0104 | .0102 | .0099 | .0096 | .0094 | .0091 | .0089 | .0087 | .0084 |

–2.2 | .0139 | .0136 | .0132 | .0129 | .0125 | .0122 | .0119 | .0116 | .0113 | .0110 |

–2.1 | .0179 | .0174 | .0170 | .0166 | .0162 | .0158 | .0154 | .0150 | .0146 | .0143 |

–2.0 | .0228 | .0222 | .0217 | .0212 | .0207 | .0202 | .0197 | .0192 | .0188 | .0183 |

–1.9 | .0287 | .0281 | .0274 | .0268 | .0262 | .0256 | .0250 | .0244 | .0239 | .0233 |

–1.8 | .0359 | .0351 | .0344 | .0336 | .0329 | .0322 | .0314 | .0307 | .0301 | .0294 |

–1.7 | .0446 | .0436 | .0427 | .0418 | .0409 | .0401 | .0392 | .0384 | .0375 | .0367 |

–1.6 | .0548 | .0537 | .0526 | .0516 | .0505 | .0495 | .0485 | .0475 | .0465 | .0455 |

–1.5 | .0668 | .0655 | .0643 | .0630 | .0618 | .0606 | .0594 | .0582 | .0571 | .0559 |

–1.4 | .0808 | .0793 | .0778 | .0764 | .0749 | .0735 | .0721 | .0708 | .0694 | .0681 |

–1.3 | .0968 | .0951 | .0934 | .0918 | .0901 | .0885 | .0869 | .0853 | .0838 | .0823 |

–1.2 | .1151 | .1131 | .1112 | .1093 | .1075 | .1056 | .1038 | .1020 | .1003 | .0985 |

–1.1 | .1357 | .1335 | .1314 | .1292 | .1271 | .1251 | .1230 | .1210 | .1190 | .1170 |

–1.0 | .1587 | .1562 | .1539 | .1515 | .1492 | .1469 | .1446 | .1423 | .1401 | .1379 |

–0.9 | .1841 | .1814 | .1788 | .1762 | .1736 | .1711 | .1685 | .1660 | .1635 | .1611 |

–0.8 | .2119 | .2090 | .2061 | .2033 | .2005 | .1977 | .1949 | .1922 | .1894 | .1867 |

–0.7 | .2420 | .2389 | .2358 | .2327 | .2296 | .2266 | .2236 | .2206 | .2177 | .2148 |

–0.6 | .2743 | .2709 | .2676 | .2643 | .2611 | .2578 | .2546 | .2514 | .2483 | .2451 |

–0.5 | .3085 | .3050 | .3015 | .2981 | .2946 | .2912 | .2877 | .2843 | .2810 | .2776 |

–0.4 | .3446 | .3409 | .3372 | .3336 | .3300 | .3264 | .3228 | .3192 | .3156 | .3121 |

–0.3 | .3821 | .3783 | .3745 | .3707 | .3669 | .3632 | .3594 | .3557 | .3520 | .3483 |

–0.2 | .4207 | .4168 | .4129 | .4090 | .4052 | .4013 | .3974 | .3936 | .3897 | .3859 |

–0.1 | .4602 | .4562 | .4522 | .4483 | .4443 | .4404 | .4364 | .4325 | .4286 | .4247 |

–0.0 | .5000 | .4960 | .4920 | .4880 | .4840 | .4801 | .4761 | .4721 | .4681 | .4641 |

0.1 | .5398 | .5438 | .5478 | .5517 | .5557 | .5596 | .5636 | .5675 | .5714 | .5753 |

0.2 | .5793 | .5832 | .5871 | .5910 | .5948 | .5987 | .6026 | .6064 | .6103 | .6141 |

0.3 | .6179 | .6217 | .6255 | .6293 | .6331 | .6368 | .6406 | .6443 | .6480 | .6517 |

0.4 | .6554 | .6591 | .6628 | .6664 | .6700 | .6736 | .6772 | .6808 | .6844 | .6879 |

0.5 | .6915 | .6950 | .6985 | .7019 | .7054 | .7088 | .7123 | .7157 | .7190 | .7224 |

0.6 | .7257 | .7291 | .7324 | .7357 | .7389 | .7422 | .7454 | .7486 | .7517 | .7549 |

0.7 | .7580 | .7611 | .7642 | .7673 | .7704 | .7734 | .7764 | .7794 | .7823 | .7852 |

0.8 | .7881 | .7910 | .7939 | .7967 | .7995 | .8023 | .8051 | .8078 | .8106 | .8133 |

0.9 | .8159 | .8186 | .8212 | .8238 | .8264 | .8289 | .8315 | .8340 | .8365 | .8389 |

1.0 | .8413 | .8438 | .8461 | .8485 | .8508 | .8531 | .8554 | .8577 | .8599 | .8621 |

Consider a population with normal distribution that has mean \(3\) and standard deviation \(4\). What is the probability that a value selected at random will be negative? What about positive?

A negative number is any number less than \(0\), so the first step is to find the \(z\)-score associated to \(0\). That is \(\frac{0 - 3}{4} = -0.75\). The value in the table associated to a value of \(-0.75\) is \(0.2266\), so there is a \(\color \red \text{22.66%} \) probability that the value will be negative. There is a \(1 - 0.2266 = 0.7734\), or \(\color \red \text{77.34%}\) probability of it being positive. \(_\square\)

## Formal Definition and Derivation

The normal distribution with mean \(\mu\) and variance \(\sigma^2\) is denoted \(\mathcal{N}(\mu, \sigma^2)\). Its probability density function is \[p_{\mu, \sigma^2} (x) = \frac{1}{\sigma \sqrt{2\pi}} e^{-\frac{(x-\mu)^2}{2\sigma^2}}.\] There is no closed form expression for the cumulative density function.

If \(X_1\) and \(X_2\) are independent normal random variables, with \(X_1 \sim \mathcal{N}(\mu_1, \sigma_1^2)\) and \(X_2 \sim \mathcal{N}(\mu_2, \sigma_2^2)\), then \(aX_1 \pm bX_2 \sim \mathcal{N}(a\mu_1 \pm b\mu_2, a^2\sigma_1^2 + b^2\sigma_2^2)\).

\[\] The bell curve is a probability density curve of binary systems. Then the probability at a some displacement from the medium is \[P(n, k) = \left( \begin{matrix} n \\ k \end{matrix} \right) {2}^{-n}= \frac{n!}{\big(\frac{1}{2}n + k\big)!\, \big(\frac{1}{2}n - k\big)!\, {2}^{n}}.\]

Using the Stirling's approximation and treating \(k = \frac{\sigma}{2}\), we have

\[P(n, \sigma) \sim {\left(\frac{n}{2\pi} \right)}^{\frac{1}{2}} {\left(\frac{n}{2}\right)}^{n} {\left(\frac{{n}^{2} - {\sigma}^{2}}{4}\right)}^{-\frac{1}{2}(n+1)}{\left(\frac{n + \sigma}{n-\sigma}\right)}^{\frac{-\sigma}{2}}.\]

For \(n>>\sigma \), \(\frac{n + \sigma}{n-\sigma} \sim 1+\frac{2\sigma}{n}\); hence, for large \(n\) \[P(n, \sigma) \sim {\left(\frac{n}{2\pi} \right)}^{\frac{1}{2}} {\left(1- \frac{{\sigma}^{2}}{{n}^{2}}\right)}^{-\frac{1}{2}(n+1)}{\left(1+\frac{2\sigma}{n}\right)}^{\frac{-\sigma}{2}}.\]

Taking the logarithm yields \[\ln\big(P(n,\sigma)\big) \sim \frac{1}{2}\ln \left (\frac{2}{\pi n}\right) - \frac{1}{2}(n+1)\ln \left (1- \frac{{\sigma}^{2}}{{n}^{2}}\right) - \frac{\sigma}{2}\ln \left (1+\frac{2\sigma}{n}\right).\]

For small \(x\), \(\ln(1+x) \approx x\); subsequently, \[\ln\big(P(n,\sigma)\big) \sim \frac{1}{2}\ln \left (\frac{2}{\pi n}\right) - \frac{1}{2}(n+1) \left (-\frac{{\sigma}^{2}}{{n}^{2}}\right) - \frac{\sigma}{2} \left (\frac{2\sigma}{n}\right)\] or \[\ln(P(n,\sigma)) \sim \frac{1}{2}\ln \left (\frac{2}{\pi n}\right) + \frac{{\sigma}^{2}}{{n}^{2}} - \frac{{\sigma}^{2}}{2n}.\]

Since \(\frac{{\sigma}^{2}}{{n}^{2}}\) vanishes faster than \(\frac{{\sigma}^{2}}{2n}\) for very large \(n\), we arrive at the result

\[P(n, \sigma) = {\left(\frac{2}{\pi n} \right)}^{\frac{1}{2}} {e}^{\frac{-{\sigma}^{2}}{2n}}.\]

## See Also

**Cite as:**Normal Distribution.

*Brilliant.org*. Retrieved from https://brilliant.org/wiki/normal-distribution/