ΣCALCULATORWizard

σ Standard Deviation Calculator

Sample SD, population SD, complete descriptive statistics, per-value z-scores, and side-by-side dataset comparison — with full deviation tables and step-by-step solutions.

Try:
Enter your data values separate by commas, spaces, or new lines
0 values Invalid values found

Sample vs. Population Standard Deviation

The most important distinction in standard deviation is whether your data represents a sample (a subset drawn from a larger group) or a population (every single member of the group you care about). The formulas differ in one critical place: what you divide by.

Population SD (σ) = √[ Σ(x − μ)² / N ]
Sample SD (s) = √[ Σ(x − x̅)² / (n − 1) ]

The sample formula divides by n−1 (called Bessel's correction). Why? When you compute the mean from sample data, you've already used one “degree of freedom” — the deviations from the sample mean are forced to sum to zero, so they carry only n−1 independent pieces of information. Dividing by n−1 instead of n corrects for this bias and makes the sample variance an unbiased estimator of the true population variance. As n grows large, the difference becomes negligible.

When to use which: If your data is the complete population (every student in a single class, all products in one batch, all members of a specific group) use population SD. If your data is a random sample and you want to infer about the broader population (1,000 survey respondents representing all voters, 50 test items representing all possible items) use sample SD.

The Full Deviation Table Explained

The deviation table is the most transparent way to see exactly how standard deviation is computed. Each row shows one data point and its contribution to the final result. The columns are: the original value x, the deviation from mean (x − x̅), the squared deviation (x − x̅)², and optionally the z-score. The final row sums the squared deviations — that sum divided by n or n−1, then square-rooted, gives the standard deviation.

Values with large squared deviations have outsized influence on SD. One extreme outlier can dramatically inflate the standard deviation, which is why it's sensitive to outliers in a way that median-based measures (like IQR) are not.

Complete Descriptive Statistics Reference

StatisticFormulaWhat it measures
Mean (μ or x̅)Σx / nAverage; center of mass of data
MedianMiddle value when sortedCenter value; robust to outliers
ModeMost frequent value(s)Most common observation
RangeMax − MinTotal spread; sensitive to outliers
Variance (s²)Σ(x − x̅)² / (n−1)Average squared deviation; in units²
Standard Deviation√VarianceTypical distance from mean; in original units
Q1 (25th pct)Median of lower halfValue below which 25% of data falls
Q3 (75th pct)Median of upper halfValue below which 75% of data falls
IQRQ3 − Q1Middle 50% spread; very robust to outliers
CV (%)(s / x̅) × 100Relative variability; unitless
SEMs / √nPrecision of mean estimate
SkewnessStandardized 3rd moment>0: right tail; <0: left tail
KurtosisExcess 4th moment>0: heavy tails; <0: light tails vs normal

The 68–95–99.7 Empirical Rule

For a normally distributed dataset, approximately 68% of values fall within 1 standard deviation of the mean, 95% within 2 standard deviations, and 99.7% within 3 standard deviations. This “empirical rule” is the foundation of quality control (Six Sigma tolerances), grading on a curve, outlier detection, and confidence interval intuition.

A value more than 2 SD from the mean is unusual (<5% expected). A value more than 3 SD away is very rare (<0.3%). This rule only applies strictly to normal distributions; skewed data may have very different percentages. This calculator shows this distribution visually for every dataset so you can immediately see whether your data has unusual spread or outliers.

Z-Scores and Standardization

A z-score measures how many standard deviations a value is from the mean: z = (x − x̅) / s. A z-score of +2.0 means the value is 2 standard deviations above average; −1.5 means 1.5 SDs below. Z-scores let you compare values from different datasets on the same scale — is a score of 85 on one test better or worse than 70 on a harder test?

Z-scores are also used for outlier detection: values with |z| > 2 are mild outliers, |z| > 3 are extreme outliers. The calculator flags these automatically in the z-score table.

Coefficient of Variation (CV)

Standard deviation is measured in the same units as your data, making it impossible to compare variability across datasets with different units or scales. The CV solves this: CV = (s / x̅) × 100%. A height dataset with SD = 10 cm and mean = 170 cm has CV = 5.9% (low variability). A price dataset with SD = $500 and mean = $1000 has CV = 50% (high variability). CV works best when data is positive and the mean is meaningfully non-zero.

Skewness and Kurtosis

Skewness measures asymmetry. Positive skewness means a long right tail (many low values, a few very high ones) — typical of income distributions. Negative skewness means a long left tail. A perfectly symmetric distribution has skewness = 0. Values beyond ±1 are generally considered notably skewed; beyond ±2, substantially skewed.

Excess kurtosis (what this calculator reports, also called Fisher kurtosis) measures tail heaviness relative to a normal distribution. Kurtosis = 0 means normal tails (mesokurtic). Positive kurtosis = heavier tails, more extreme values than expected (leptokurtic). Negative = lighter tails, fewer extremes (platykurtic). Financial return data typically shows high positive kurtosis — rare but very large gains and losses occur more often than a normal distribution would predict.

Frequently Asked Questions

Should I use sample or population standard deviation?
Use population SD (σ) when your dataset IS the entire population you care about — all 30 students in a specific class, all 500 products from one machine run. Use sample SD (s) when your dataset is a subset of a larger population you want to make inferences about — 100 survey respondents out of millions, 20 measurements taken from an ongoing process. When in doubt: if you collected a subset and want to generalize, use sample SD.
What does standard deviation actually tell me?
It tells you the typical distance any random data point is from the mean. If exam scores have a mean of 75 and SD of 10, a typical student scores somewhere between 65 and 85. A small SD means data clusters tightly around the mean (consistent). A large SD means data is widely spread (variable). Two datasets can have the same mean but very different SDs — SD is what distinguishes them.
Why does standard deviation use squared deviations instead of absolute deviations?
Squaring deviations has three advantages: it makes all deviations positive (eliminating the cancel-out problem), it penalizes large deviations more heavily than small ones (making SD sensitive to outliers), and it produces mathematically tractable properties — variance adds across independent variables, which absolute deviations don't. The mean absolute deviation (MAD) is an alternative that's more robust to outliers but lacks these algebraic properties.
What is a "good" or "bad" standard deviation?
There's no universal answer — it entirely depends on context. For manufacturing tolerances, a small SD is good (consistent output). For investment returns, some SD is necessary (no SD = no return). The Coefficient of Variation (CV) provides more context: CV < 15% is generally considered low variability, 15–30% moderate, >30% high — but these are rough guidelines, not universal rules.
How do outliers affect standard deviation?
Dramatically. Because deviations are squared, one value far from the mean contributes far more to SD than values close to the mean. Dataset [1,2,3,4,5] has SD ≈ 1.58. Add one outlier: [1,2,3,4,5,50] and SD jumps to ≈ 17.9. This is why SD is called "not robust" to outliers. The IQR (interquartile range) is the robust alternative — it ignores values outside the middle 50% entirely.
What does a negative z-score mean?
A negative z-score means the value is below the mean. z = −1.5 means the value is 1.5 standard deviations below average. Z-scores can be any real number — positive (above mean), negative (below mean), or zero (exactly at the mean). There's nothing inherently bad about a negative z-score; it just indicates position relative to the group average.
What's the difference between standard deviation and standard error?
Standard deviation (SD) measures the spread of individual data points in your sample. Standard error of the mean (SEM = SD/√n) measures the precision of your estimate of the population mean. As you collect more data (larger n), SEM shrinks but SD stays roughly constant. SD answers "how variable is the data?" SEM answers "how precisely do I know the mean?" Error bars in scientific papers typically show SEM, not SD.
Can standard deviation be zero or negative?
Zero, yes — if every value in the dataset is identical, all deviations are zero, so SD = 0. This means no variability whatsoever. Negative, no — SD is a square root, and variance (the thing being square-rooted) is a sum of squared values which is always ≥ 0. A negative standard deviation is mathematically impossible, so any result suggesting otherwise indicates a calculation error.