0% found this document useful (0 votes)
94 views

Standard Deviation and Standard Error

1) The standard deviation (SD) measures the dispersion of data points around the mean and is used to describe variability within a sample. 2) The standard error of the mean (SEM) measures how far the sample mean is likely to be from the true population mean and describes variability between samples. 3) Researchers often confuse SD and SEM. The SD is reported to describe variability within a sample, while the SEM is used to create confidence intervals around the mean between samples.

Uploaded by

deepshikha.s
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
94 views

Standard Deviation and Standard Error

1) The standard deviation (SD) measures the dispersion of data points around the mean and is used to describe variability within a sample. 2) The standard error of the mean (SEM) measures how far the sample mean is likely to be from the true population mean and describes variability between samples. 3) Researchers often confuse SD and SEM. The SD is reported to describe variability within a sample, while the SEM is used to create confidence intervals around the mean between samples.

Uploaded by

deepshikha.s
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

Teaching Mathematics and Its Applications (2010) 29, 108^112

doi:10.1093/teamat/hrq003 Advance Access publication 12 March 2010

A note on standard deviation and


standard error
HOSSEIN HASSANI,MANSOUREH GHODSI AND GARETH HOWELL
Group of Statistics, Cardiff School of Mathematics, Cardiff University, Senghennydd Road,
Cardiff CF24 4AG, UK

Downloaded from http://teamat.oxfordjournals.org at Cardiff University on June 4, 2010



Email: hassanih@cf.ac.uk

[Submitted October 2009; accepted January 2010]

Many students confuse the standard deviation and standard error of the mean and
are unsure which, if either, to use in presenting data. In this article, we endeavour to
address these questions and cover some related ambiguities about these quantities.

1. Introduction
When reporting data in research papers, authors often use descriptive statistics to describe the basic
features of the data sets. Descriptive statistics provide simple summaries about the sample. In general,
descriptive statistics include the statistical measures of central tendency and dispersion. The standard
deviation (SD) is one of the common measures of dispersion; the SD is an index of variability of the
original data points and is reported in most studies (Streiner, 1996). The precise term ‘standard devi-
ation’ was first used in writing by Karl Pearson in 1894 following his usage in lectures on the topic
(Dodge, 2003).
The SD of the sampling distribution associated with the estimation method is the standard error of
estimation (Everitt, 2003). The standard error of the mean (SEM) is the SD of the sample mean
estimate of a population mean. It can also be viewed as the SD of the error in the sample mean
relative to the true mean, since the sample mean is an unbiased estimator of the population mean. The
SE reflects the variability of the mean values, as if the study were repeated a large number of times.
A big question often posed for beginners in statistics is: ‘when should we report the standard
deviation and when should we report the standard error?’ Students get these confused easily and
ask how SD and SE are correlated, and what is the difference between them? Does SD describe the
true distance of measurements from the mean? Is the standard error just an estimate of this?
Moreover, authors in their research reports often use SE to describe the variability of their
sample (Nagele, 2003), which may seem to indicate that this problem is not simply confined
to beginners and students in statistics. For example, Nagele (2003) showed that even in some
leading journals, there are a significant number of published articles that misuse the SE in descrip-
tive statistics, which may be misinterpreted as showing the variability within the study sample.
In this article, we endeavour to address these questions and cover some related ambiguities about
SD and SE.

ß The Author 2010. Published by Oxford University Press on behalf of The Institute of Mathematics and its Applications.
All rights reserved. For permissions, please email: journals.permissions@oxfordjournals.org
H. HASSANI ET AL. 109

Section 2 discusses the SD, its construction, the structure of its formula and related features.
Appropriate examples are included for further clarification. The standard error was considered in
Section 3. The relation between SE and SD and the use of SE in confidence intervals is also considered
in Section 3. Finally, Section 4 provides some summaries regarding to the comparison between these
two entities.

2. SD

Downloaded from http://teamat.oxfordjournals.org at Cardiff University on June 4, 2010


The standard deviation, abbreviated variously as S.D., SD, or S is an index of how closely the indi-
vidual data points cluster around the mean. If we call each data point Xi , i ¼ 1 . . . ; N (N is the
population size) and X the sample
P mean, then it may seem that an index of the dispersion of the
data points would simply be Ni¼1 ðXi  XÞ,
 which means to sum how much each value of Xi deviates
 In other words, an index of dispersion would be:
from X.
X
N
ðIndividual data point  Mean of the data pointsÞ:
i¼1

Logical as this may seem, it has two drawbacks. The first difficulty is that the answer will be zero—not
just in this situation, but in every case. By definition, the sum of the values above the mean is always
equal to the sum of the values below it, and thus they will cancel each other out. Let us first consider
the following example. Assume observations 10, 12, 20, 8 and 15 are our sample. The mean of this
sample is X ¼ 13; however, the sum in question is
X
N
 ¼ ð10  13Þ þ ð12  13Þ þ ð20  13Þ þ ð8  13Þ þ ð15  13Þ ¼ 0:
ðXi  XÞ
i¼1

We can get around this problem by taking the absolute value of each difference (i.e. we can ignore the
sign of the differences whenever they are negative), but for a number of reasons, statisticians do not
like to use absolute numbers. Another way to eliminate negative values is to square them, since the
square of any number—negative or positive—is always positive. So, what we now have is
X
N
 2:
ðXi  XÞ ð1Þ
i¼1

The second problem is that the result of this equation will increase as we successively add more
objects; this is of little use particularly for large data sets. Let us imagine that we have a sample of 50
values, for which the result of (1) is 20. If we now add another 50 subjects who look exactly the same,
it makes intuitive sense that the dispersion of these 100 points should stay the same. Yet, the formula
as it now reads can result only in a larger sum as we add more data points. So, we can compensate
for this by dividing by the number of subjects, N, so that the equation now reads
PN
 2
i¼1 ðXi  XÞ
: ð2Þ
N
Suppose we are interested in the average height of all men in a population. To find out the true mean,
we would have to observe (i.e. measure) the height of each man in the population. Of course, in studies
we are unable to observe everyone in a population, because of the large size or the difficulty and
expense of reaching every member. Therefore, we use samples that can give us reasonable estimates.
110 A NOTE ON SD AND SE

Assume we found that the mean of a sample is X ¼ 69:7 cm and its dispersion according to (2) is
9 cm2 . This would seem to engender another set of questions: why is the dispersion measure in square
centimetres, and what can we do to interpret this number? Fortunately, these issues are easily
explained: naturally the units become cm2, since each length ðXi  XÞ  is squared, hence we need to
take the square root of the final solution given by (2) to return to the correct dimension. This now
provides us with an answer which is much easier to interpret as a measure of dispersion.
There is, however, one last problem when dealing with the SD of a sample such as the one described
above. The result of the formula as it stands thus far produces a biased estimate—i.e. one that is
consistently either higher or (as in this case) lower than the ‘true’ value because we are limited to using

Downloaded from http://teamat.oxfordjournals.org at Cardiff University on June 4, 2010


only a sample of the population and not its entirety. This is known as a biased estimator. In fact, a clear
distinction must be made at this stage between the population SD, which divides by N, and the sample
SD, which uses the quantity N – 1.

2.1 Why do we use N – 1 and not N?


The reasoning is based on the fact that we almost always use the sample variance to obtain an estimate
of the population variance (a population being all the measurements or events of the same type that
could ever be found). Put in simple terms, the population variance is estimated from the sample mean
and from the deviation (d) of each measurement from the sample mean. If we lack any one of these
measurements (the mean or a single d value), we could calculate it from the other information. So, with
N measurements (data points) only N – 1 of them are free to vary when we know the mean; and one
could easily calculate the missing value. ‘N – 1’ is therefore the number of degrees of freedom of our
data.
In summary, since the sum (1) is necessarily zero, and there are N such points of consideration, then
we can conclude that they are not independent of each other. Now, summing just N – 1 of the
differences ðXi  XÞ gives a non-zero result, and hence the condition of independence is upheld. It
is this correction—known as Bessel’s Correction (Reichmann, 1961)—made to formula (2) that
removes bias from the sample SD.
Putting all of this together, we finally arrive at the formula for the SD for a sample, which is:
sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
PN
i¼1 ðXi  XÞ

SD ¼ : ð3Þ
N 1
Now that we have gone through all this work, what does it all mean? If we assume that the data are
normally distributed, then knowing the mean and SD tells us everything we need to know about the
distribution of observations. In any normal distribution, roughly two-thirds (actually, 68.2%) of the
observations fall between ð1SD; þ1SDÞ, and 95.4% betweenð2SD; þ2SDÞ. Therefore, the SD
tells us the distribution of individual scores around the mean. Now, let us turn our attention to the
standard error.

3. Standard error
We mentioned previously that the purpose of most studies is to estimate some population parameter,
such as the mean, the SD, a correlation or a proportion. Once we have that estimate, another question
then arises: How accurate is our estimate? This may seem an unanswerable question; if we do not
know what the population value is, how can we evaluate how close we are to it?
H. HASSANI ET AL. 111

Mere logic, however, has never stopped statisticians in the past, and it will not stop us now. What we
can do is resort to probabilities: What is the probability (P) that the true (population) mean falls within
a certain range of values? (To cite one of our mottos, ‘Statistics means you never have to say you are
certain.’). One way to answer the question is to repeat the study a few hundred times, which will give
us many estimates of the mean. We can then take the mean of these means, as well as figure out what
the distribution of means is; that is, we can get the SD of the mean values. Then, using the same table
of normal distribution as usual, we can estimate what range of values would encompass 90 or 95% of
the means. If each sample had been drawn from the population at random, we would be fairly safe in
concluding that the true mean also falls within this range 90 or 95% of the time. We assign a new name

Downloaded from http://teamat.oxfordjournals.org at Cardiff University on June 4, 2010


to the SD of the means: we call it the standard error of the mean (abbreviated as SEM, or, if there is no
ambiguity that we are talking about the mean, SE).
First, let us deal with one slight problem—replicating the study a few hundred times. Nowadays, it is
hard enough to get money to do a study once, much less replicate it this many times (even assuming
you would actually want to spend the rest of your life doing the same study over and over). Ever
helpful, statisticians have figured out a way to determine the SE based on the results of a single study.
Let us approach this first from an intuitive standpoint: what would make us more or less confident that
an estimate of the population mean, based on our study, is accurate? One obvious thing would be the
size of the study; the larger the sample size, N, the less chance that one or two aberrant values are
distorting the results and the more likely it is that our estimate is close to the true value.
Intuitively, one would say that some index of N should be in the denominator of SE, since the larger
N is, the smaller SE would become. Second, and for similar reasons, the smaller the variability in the
data, the more confident we are that one value (the mean) accurately reflects them. Thus, the SD
should be in the numerator: the larger it is, the larger SE will be, and we end up with the equation:
SD
SE ¼ pffiffiffiffi ð4Þ
N
pffiffiffiffi
Another question: why does the denominator read N instead of just N?
This is since we are really dividing the variance, which is SD2 , by N, but we end up again with
squared units, so we take the square root of everything to again reflect to the correct dimension.
To recap, the SD thus reflects the variability of individual data points, and the SE is the variability of
means. Standard errors provide simple measures of uncertainty in a value and are often used because:
(1) If the standard error of several individual quantities is known, then the standard error of some
function of the quantities can be easily calculated in many cases;
(2) Where the probability distribution of the value is known, it can be used to calculate an exact
confidence interval;
(3) Where the probability distribution is unknown, relationships like Chebyshev’s or the
Vysochanskij–Petunin (Vysochanskij & Petunin, 1980) inequality can be used to calculate a
conservative confidence interval;
(4) As the sample size tends to infinity the central limit theorem guarantees that the sampling distri-
bution of the mean is asymptotically normal.

4. Summary
pffiffiffiffi
Although the SD and the SE are related ðSE ¼ SD= N ), they give two very different types of infor-
mation (Carlin & Doyle, 2000). Moreover, it is easy to be confused about the difference between SD
112 A NOTE ON SD AND SE

and SE. Whereas the SD estimates the variability in the study sample, the SE estimates the precision
and uncertainty of how the study sample represents the underlying population (Webster & Merry,
1997). In other words, the SD tells us the distribution of individual data points around the mean, and
the SE informs us how precise our estimate of the mean is (Streiner, 1996). In summary, we can
mention the following differences:
(1) The SD quantifies scatter—how much the values vary from one another.
(2) The SE quantifies how accurately you know the true mean of the population.
(3) The SE, by definition, is always smaller than the SD.
(4) The SE gets smaller as your samples get larger. This makes sense, because the mean of a large

Downloaded from http://teamat.oxfordjournals.org at Cardiff University on June 4, 2010


sample is likely to be closer to the true population mean than is the mean of a small sample. With
a huge sample, you will know the value of the mean with a lot of precision even if the data vary
a lot.
(5) The SD does not change predictably as you acquire more data. The SD quantifies the scatter of the
data, and increasing the size of the sample does not change the scatter. The SD might go up or it
might go down; you cannot predict. In practice, the SD will stay the same as sample size gets
larger.

REFERENCES
CARLIN, J. B. & DOYLE, L. W. (2000) Basic concepts of statistical reasoning: standard errors and confidence
intervals. J Paediatr Child Health, 36, 502–505.
DODGE, Y. (2003) The Oxford Dictionary of Statistical Terms, Oxford: Oxford University Press. ISBN
0-19-920613-9.
EVERITT, B. S. (2003) The Cambridge Dictionary of Statistics, CUP. ISBN 0-521-81099-x.
NAGELE, P. (2003) Misuse of standard error of the mean (SEM) when reporting variability of a sample.
A critical evaluation of four anaesthesia. Br J Anaesthesiol, 90, 514–516.
REICHMANN, J. (1961) Use and Abuse of Statistics, London: Methuen. Reprinted 1964–1970 by Pelican.
Appendix 8.
STREINER, D. L. (1996) Maintaining standards: differences between the standard deviation and standard error,
and when to use each. Can. J. Psychiatry, 41, 498–502.
VYSOCHANSKIJ, D. F. & PETUNIN, Y. I. (1980) Justification of the 3s rule for unimodal distributions. Theory
Probab Math Stat, 21, 25–36.
WEBSTER, C. S. & MERRY, A. F. (1997) The standard deviation and the standard error of the mean. Anaesthesia,
52, 183.

Hossein Hassani was born on 22 October 1976 in Tehran, Iran. He graduated with a Bsc and Msc in
Statistics (IRAN) and Phd in Statistics at the Cardiff School of Mathematics in December 2009. He has been
involved in teaching Statistics since 2004 at different levels, and also various subjects. He has been working
as a lecturer in Statistics at Cardiff School of Medicine since December 2009. He has published several
papers in high scholarly journals (some of them are the top journals in their fields). He has also received
several international prizes and grants for his research during his PhD study.
Mansoureh Ghodsi has been working towards her doctorate under the stewardship of the Statistics Group
at the Cardiff University since late 2006 and is due to submit for her PhD in 2011.
Gareth Howell has been working towards his doctorate under the stewardship of the Statistics Group at the
Cardiff University since late 2006 and is due to complete it very shortly.

View publication stats

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy