The normal distribution, also known as the Gaussian distribution, is a theoretical continuous distribution of a random variable - and is mathematically defined by several formulae. For non-mathematicians, a qualitative description of its properties may be more useful. Show The normal distribution was so named because it was thought to be the natural or normal distribution for any continuous variable to follow. We now know that, in biology at least, that is not necessarily the case. But in statistics the distribution remains extremely important because it more-or-less describes the random variation of sample means - and many statistics that behave as means. {Fig. 1} Whilst many people visualise a normal distribution as a 'bell-shaped' curve, for a critical appraisal you need to define its properties much more clearly and quantitatively. Let us begin by stating the properties of the distribution.
Let us consider how varying these population parameters affects the appearance of the distribution.
The population's locationBecause the normal distribution is smooth and symmetrical, the mean, median, and mode of any normal population are identical. The graph below shows the distribution of 3 normal populations, whose only difference is their location. {Fig. 2} The population mean is usually defined as the mean of all the values in that population - or μ. More explicitly, if we call our population X, μx would be the population mean. In contrast, the mean of a sample of that population, is only an estimate of the 'true' population mean, and the two are only the same on average.Another way of defining the population mean is in terms of the average result of randomly sampling that population. For example, if x is an observation of population X then, if we took an infinitely large sample its mean would be Σx/∞ - which causes a few headaches! To avoid this dilemma, we say that, if we repeatedly sample a population, we would expect the average value of x, E(x) to be identical to the population mean, μx. In other words, μx = E(x) = E( ) - or various other mathematical formulae, depending upon the context.If the probability of observing a value was unrelated to the value being observed the distribution would be uniform. However, if you sample a normal population at random, the most commonly observed values are closest to the population mean.
The population's dispersionWith a normal population, for mathematical reasons, the dispersion is usually defined in terms of the root mean squared deviation of its observations about their population mean - in other words the population standard deviation, σ. The graph below shows the distribution of 3 normal populations, whose only difference is their dispersion. {Fig. 3} By convention, the standard deviation of a population called Y is generally represented by the Greek letter s - in other words σy - or just σ. The standard deviation of a sample of that population may be written as sy, or just s. Aside from their mean and standard deviation, every normal population is identical. Therefore, if you rescale normal populations to allow for these two parameters, every normal population is completely identical. The commonest way to rescale a normal population is to subtract the mean from each observation, and divide by the standard deviation. This will produce a standard normal population, which has a mean of zero and a standard deviation of one. This is mathematically the simplest of all - and, because is so useful, the standard normal distribution has its own special symbols and terminology.
Central limitAs we said above, the main reason why the normal distribution is so important in statistics is that many sample statistics, including the mean, tend towards a normal distribution, irrespective of the population distribution. The way in which a statistic's normal tendency depends upon sample size is described by what is known as central limit theorem. Non-mathematically, there are three factors which determine how large a sample you need in order to assume a statistic is approximately normal.
Another reason the normal distribution is so popular is because its properties are well known - at least to mathematicians. In particular, there are various formulae which estimate the proportion of a normal population in a defined interval. These are known as probability functions. Related topics :Skewness and kurtosis Chauvenet's criterion for identifying outliersThe log normal distribution Gaussian smoothing
For the majority of the remainder of this class, we’ll be focusing on variables that have a (roughly) normal distribution. For example, data sets consisting of physical measurements (heights, weights, lengths of bones, and so on) for adults of the same species and sex often follow a similar pattern: most individuals are clumped around the average or mean of the population, with numbers decreasing the farther values are from the average in either direction. The shape of any normal curve is a single-peaked, symmetric distribution that is bell-shaped. A normally distributed random variable, or a variable with a normal probability distribution, is a continuous random variable that has a relative frequency histogram in the shape of a normal curve. This curve is also called the normal density curve. The actual functional notation for creating the normal curve is quite complex: where μ and σ are the mean and standard deviation of the population of data. What this formula tells us is that any mean μ and standard deviation σ completely define a unique normal curve. Recall that μ tells us the “center” of the peak while σ describes the overall “fatness” of the data set. A small σ value indicates a tall, skinny data set, while a larger value of σ results in a shorter, more spread out data set. Each normal distribution is indicated by the symbols N(μ,σ) . For example, the normal distribution N(0,1) is called the standard normal distribution, and it has a mean of 0 and a standard deviation of 1. Properties of a Normal Distribution
A graphical representation of the Normal Distribution curve below: Because there are an infinite number of possibilities for µ and σ, there are an infinite number of normal curves. In order to determine probabilities for each normally distributed random variable, we would have to perform separate probability calculations for each normal distribution. One amazing fact about any normal distribution is called the 68-95-99.7 Rule, or more concisely, the empirical rule. This rule states that:
A graphical representation of the empirical rule is shown in the following figure: Image from: http://2.bp.blogspot.com/-J2YOCi9-1Tg/U95XGRQBS-I/AAAAAAAABKQ/y5vD4qMSJb4/s1600/stdeviation.png Example:Suppose a variable has mean μ = 17 and standard deviation σ = 3.4. Then, according to the empirical rule:
The results from the third bullet point illustrate how a data value of, say, 2.1 (which is less than 6.8) or a data value of, say, 33.2 (a value greater than 27.2) would both be very unusual, since almost all data values should lie between 6.8 and 27.2. Back to the Standard Normal CurveAll normal distributions, regardless of their mean and standard deviation, share the Empirical Rule. With some very simple mathematics, we can “transform” any normal distribution into the standard normal distribution. This is called a z-transform. Using the z-transformation, any data set that is normally distributed can be converted to the same standard normal distribution by the conversion: where X is the normally distributed random variable, and Z is a random variable following the standard normal distribution. Notice when X = μ that Z = (μ – μ)/σ = 0, which explains how Z transforms our mean to 0. Properties of the Standard Normal Distribution
|