The Central Limit Theorem
In the practice of statistics, most problems involving a significance test (z or t test), finding a probability, or the determination of a confidence interval requires the usage of normal approximations. Most populations have roughly normal distributions that allow for a random sample to be taken and tested without caution. However, there are some populations that just don’t fit the description of “normal”. They don’t follow a precise bell curve, and the lack of normality could leave a statistician wondering how they could possibly determine if normal approximations are appropriate. It is at this point that one of the most important theorems in statistical mathematics comes into play. The Central Limit Theorem, which was first introduced by Pierre-Simon Laplace in the late 18th century, states that as long as other important conditions are satisfied, a sample will become more normal as its size increases. An example of this is shown below. [pic]
As you can see, as “N” increased, the distribution began to appear more like a bell curve, which is just another way of saying “normal”.
Some rules have been applied to the Central Limit Theorem throughout its history that have contributed to it becoming one of the most consistently accurate mathematical theorems. For example, there are the previously mentioned ‘conditions’ that need to be met. In order to utilize the theorem, the sample must be taken randomly, pulled from the population, and it must show independence. The basic way to determine independence is through the 10% condition, which says that a sample is independent if it makes up less than 10% of the overall population. There is also the issue of deciding just how “large” the sample must be in order to be considered “large” enough to assume normality. For the most part, as long as the sample size (N) is greater than 30, the Central Limit Theorem can be applied. If N is between 10...
Please join StudyMode to read the full document