Jump to content

Standard error

From Wikipedia, the free encyclopedia
For a value that is sampled with an unbiasednormally distributederror, the above depicts the proportion of samples that would fall between 0, 1, 2, and 3 standard deviations above and below the actual value.

Thestandard error(SE)[1]of astatistic(usually an estimate of aparameter) is thestandard deviationof itssampling distribution[2]or an estimate of that standard deviation. If the statistic is the sample mean, it is called thestandard error of the mean(SEM).[1]The standard error is a key ingredient in producingconfidence intervals.[3]

Thesampling distributionof a mean is generated by repeated sampling from the same population and recording of the sample means obtained. This forms a distribution of different means, and this distribution has its ownmeanandvariance.Mathematically, the variance of the sampling mean distribution obtained is equal to the variance of the population divided by the sample size. This is because as the sample size increases, sample means cluster more closely around the population mean.

Therefore, the relationship between the standard error of the mean and the standard deviation is such that, for a given sample size, the standard error of the mean equals the standard deviation divided by thesquare rootof the sample size.[1]In other words, the standard error of the mean is a measure of the dispersion of sample means around the population mean.

Inregression analysis,the term "standard error" refers either to the square root of thereduced chi-squared statisticor the standard error for a particular regression coefficient (as used in, say,confidence intervals).

Standard error of the sample mean

[edit]

Exact value

[edit]

Suppose a statistically independent sample ofobservationsis taken from astatistical populationwith astandard deviationof.The mean value calculated from the sample,,will have an associatedstandard error on the mean,,given by:[1]

Practically this tells us that when trying to estimate the value of a population mean, due to the factor,reducing the error on the estimate by a factor of two requires acquiring four times as many observations in the sample; reducing it by a factor of ten requires a hundred times as many observations.

Estimate

[edit]

The standard deviationof the population being sampled is seldom known. Therefore, the standard error of the mean is usually estimated by replacingwith thesample standard deviationinstead:

As this is only anestimatorfor the true "standard error", it is common to see other notations here such as:

A common source of confusion occurs when failing to distinguish clearly between:

  • the standard deviation of thepopulation(),
  • the standard deviation of thesample(),
  • the standard deviation of themeanitself (,which is the standard error), and
  • theestimatorof the standard deviation of the mean (,which is the most often calculated quantity, and is also often colloquially called thestandard error).

Accuracy of the estimator

[edit]

When the sample size is small, using the standard deviation of the sample instead of the true standard deviation of the population will tend to systematically underestimate the population standard deviation, and therefore also the standard error. Withn= 2, the underestimate is about 25%, but forn= 6, the underestimate is only 5%. Gurland and Tripathi (1971) provide a correction and equation for this effect.[4]Sokal and Rohlf (1981) give an equation of the correction factor for small samples ofn< 20.[5]Seeunbiased estimation of standard deviationfor further discussion.

Derivation

[edit]

The standard error on the mean may be derived from thevarianceof a sum of independent random variables,[6]given thedefinitionof variance and somepropertiesthereof. Ifis a sample ofindependent observations from a population with meanand standard deviation,then we can define the total which due to theBienaymé formula,will have variance where we've approximated the standard deviations, i.e., the uncertainties, of the measurements themselves with the best value for the standard deviation of the population. The mean of these measurementsis given by

The variance of the mean is then

The standard error is, by definition, the standard deviation ofwhich is the square root of the variance:

For correlated random variables the sample variance needs to be computed according to theMarkov chain central limit theorem.

Independent and identically distributed random variables with random sample size

[edit]

There are cases when a sample is taken without knowing, in advance, how many observations will be acceptable according to some criterion. In such cases, the sample sizeis a random variable whose variation adds to the variation ofsuch that,[7] which follows from thelaw of total variance.

Ifhas aPoisson distribution,thenwith estimator.Hence the estimator ofbecomes,leading the following formula for standard error: (since the standard deviation is the square root of the variance).

Student approximation whenσvalue is unknown

[edit]

In many practical applications, the true value ofσis unknown. As a result, we need to use a distribution that takes into account that spread of possibleσ's. When the true underlying distribution is known to be Gaussian, although with unknown σ, then the resulting estimated distribution follows the Student t-distribution. The standard error is the standard deviation of the Student t-distribution. T-distributions are slightly different from Gaussian, and vary depending on the size of the sample. Small samples are somewhat more likely to underestimate the population standard deviation and have a mean that differs from the true population mean, and the Student t-distribution accounts for the probability of these events with somewhat heavier tails compared to a Gaussian. To estimate the standard error of a Student t-distribution it is sufficient to use the sample standard deviation "s" instead ofσ,and we could use this value to calculate confidence intervals.

Note:TheStudent's probability distributionis approximated well by the Gaussian distribution when the sample size is over 100. For such samples one can use the latter distribution, which is much simpler. Also, even though the 'true' distribution of the population is unknown, assuming normality of the sampling distribution makes sense for a reasonable sample size, and under certain sampling conditions, seeCLT.If these conditions are not met, then using aBootstrap distributionto estimate the Standard Error is often a good workaround, but it can be computationally intensive.

Assumptions and usage

[edit]

An example of howis used is to make confidence intervals of the unknown population mean. If the sampling distribution isnormally distributed,the sample mean, the standard error, and thequantilesof the normal distribution can be used to calculate confidence intervals for the true population mean. The following expressions can be used to calculate the upper and lower 95% confidence limits, whereis equal to the sample mean,is equal to the standard error for the sample mean, and1.96is the approximate value of the 97.5percentilepoint of thenormal distribution:

  • Upper 95% limit =,and
  • Lower 95% limit =.

In particular, the standard error of asample statistic(such assample mean) is the actual or estimated standard deviation of the sample mean in the process by which it was generated. In other words, it is the actual or estimated standard deviation of thesampling distributionof the sample statistic. The notation for standard error can be any one of SE, SEM (for standard error ofmeasurementormean), or SE.

Standard errors provide simple measures of uncertainty in a value and are often used because:

Standard error of mean versus standard deviation

[edit]

In scientific and technical literature, experimental data are often summarized either using the mean and standard deviation of the sample data or the mean with the standard error. This often leads to confusion about their interchangeability. However, the mean and standard deviation aredescriptive statistics,whereas the standard error of the mean is descriptive of the random sampling process. The standard deviation of the sample data is a description of the variation in measurements, while the standard error of the mean is a probabilistic statement about how the sample size will provide a better bound on estimates of the population mean, in light of the central limit theorem.[8]

Put simply, thestandard errorof the sample mean is an estimate of how far the sample mean is likely to be from the population mean, whereas thestandard deviationof the sample is the degree to which individuals within the sample differ from the sample mean.[9]If the population standard deviation is finite, the standard error of the mean of the sample will tend to zero with increasing sample size, because the estimate of the population mean will improve, while the standard deviation of the sample will tend to approximate the population standard deviation as the sample size increases.

Extensions

[edit]

Finite population correction (FPC)

[edit]

The formula given above for the standard error assumes that the population is infinite. Nonetheless, it is often used for finite populations when people are interested in measuring the process that created the existing finite population (this is called ananalytic study). Though the above formula is not exactly correct when the population is finite, the difference between the finite- and infinite-population versions will be small whensampling fractionis small (e.g. a small proportion of a finite population is studied). In this case people often do not correct for the finite population, essentially treating it as an "approximately infinite" population.

If one is interested in measuring an existing finite population that will not change over time, then it is necessary to adjust for the population size (called anenumerative study). When thesampling fraction(often termedf) is large (approximately at 5% or more) in anenumerative study,the estimate of the standard error must be corrected by multiplying by a ''finite population correction'' (a.k.a.:FPC):[10] [11] which, for largeN: to account for the added precision gained by sampling close to a larger percentage of the population. The effect of the FPC is that the error becomes zero when the sample sizenis equal to the population sizeN.

This happens insurvey methodologywhen samplingwithout replacement.If sampling with replacement, then FPC does not come into play.

Correction for correlation in the sample

[edit]
Expected error in the mean ofAfor a sample ofndata points with sample bias coefficientρ.The unbiasedstandard errorplots as theρ= 0 diagonal line with log-log slope −12.

If values of the measured quantityAare not statistically independent but have been obtained from known locations in parameter spacex,an unbiased estimate of the true standard error of the mean (actually a correction on the standard deviation part) may be obtained by multiplying the calculated standard error of the sample by the factorf: where the sample bias coefficient ρ is the widely usedPrais–Winsten estimateof theautocorrelation-coefficient (a quantity between −1 and +1) for all sample point pairs. This approximate formula is for moderate to large sample sizes; the reference gives the exact formulas for any sample size, and can be applied to heavily autocorrelated time series like Wall Street stock quotes. Moreover, this formula works for positive and negative ρ alike.[12]See alsounbiased estimation of standard deviationfor more discussion.

See also

[edit]

References

[edit]
  1. ^abcdAltman, Douglas G; Bland, J Martin (2005-10-15)."Standard deviations and standard errors".BMJ: British Medical Journal.331(7521): 903.doi:10.1136/bmj.331.7521.903.ISSN0959-8138.PMC1255808.PMID16223828.
  2. ^Everitt, B. S. (2003).The Cambridge Dictionary of Statistics.Cambridge University Press.ISBN978-0-521-81099-9.
  3. ^Wooldridge, Jeffrey M. (2023)."What is a standard error? (And how should we compute it?)".Journal of Econometrics.237(2, Part A).doi:10.1016/j.jeconom.2023.105517.ISSN0304-4076.
  4. ^Gurland, J; Tripathi RC (1971). "A simple approximation for unbiased estimation of the standard deviation".American Statistician.25(4): 30–32.doi:10.2307/2682923.JSTOR2682923.
  5. ^Sokal; Rohlf (1981).Biometry: Principles and Practice of Statistics in Biological Research(2nd ed.). p.53.ISBN978-0-7167-1254-1.
  6. ^Hutchinson, T. P. (1993).Essentials of Statistical Methods, in 41 pages.Adelaide: Rumsby.ISBN978-0-646-12621-0.
  7. ^Cornell, J R; Benjamin, C A (1970).Probability, Statistics, and Decisions for Civil Engineers.NY: McGraw-Hill. pp. 178–179.ISBN0486796094.
  8. ^Barde, M. (2012)."What to use to express the variability of data: Standard deviation or standard error of mean?".Perspect. Clin. Res.3(3): 113–116.doi:10.4103/2229-3485.100662.PMC3487226.PMID23125963.
  9. ^Wassertheil-Smoller, Sylvia(1995).Biostatistics and Epidemiology: A Primer for Health Professionals(Second ed.). New York: Springer. pp. 40–43.ISBN0-387-94388-9.
  10. ^Isserlis, L. (1918)."On the value of a mean as calculated from a sample".Journal of the Royal Statistical Society.81(1): 75–81.doi:10.2307/2340569.JSTOR2340569.(Equation 1)
  11. ^Bondy, Warren; Zlot, William (1976). "The Standard Error of the Mean and the Difference Between Means for Finite Populations".The American Statistician.30(2): 96–97.doi:10.1080/00031305.1976.10479149.JSTOR2683803.(Equation 2)
  12. ^Bence, James R. (1995)."Analysis of Short Time Series: Correcting for Autocorrelation".Ecology.76(2): 628–639.doi:10.2307/1941218.JSTOR1941218.