Standard Deviation Calculator
Enter your dataset — get population SD, sample SD, variance, mean, and more instantly
2, 4, 4, 4, 5, 5, 7, 9Calculation Result
Full statistical breakdown
Calculation Summary
Step-by-Step Working
Verification — Key Stats Check
All Statistical Measures
Full Details & Interpretation
What Is Standard Deviation? A Complete Beginner's Guide
Understand standard deviation from first principles — definition, notation, formulas, and why it matters across every field
Standard deviation is a measure of how spread out the values in a dataset are relative to the mean (average). A low standard deviation means values are clustered tightly around the mean; a high standard deviation means values are widely scattered. It is the single most widely used measure of statistical dispersion in the world.
Mathematically, standard deviation is the square root of variance, and variance is the average of the squared differences from the mean. The reason we square the differences is to make all deviations positive (so negatives don't cancel positives) and to penalize large deviations more heavily than small ones.
There are two versions you must know: population standard deviation (σ), used when you have every member of the group, and sample standard deviation (s), used when you have a subset. The difference is whether you divide by N or N−1 — a correction called Bessel's correction that prevents underestimation of the true spread.
Standard Deviation Formulas — Complete Reference
Every formula you need: population, sample, variance, z-score, CV — with worked examples
| Measure | Formula | When to Use | Example |
|---|---|---|---|
| Mean (Average) | μ = Σx / N |
Central tendency of any dataset | Mean of {2,4,6}: (2+4+6)/3 = 4 |
| Population Variance | σ² = Σ(x − μ)² / N |
Full population data available | Squared deviations averaged over all N points |
| Population SD | σ = √[Σ(x − μ)² / N] |
Full population; exam grades for entire class | σ of {2,4,4,4,5,5,7,9} = 2.0 |
| Sample Variance | s² = Σ(x − x̄)² / (N−1) |
Sample data; estimate of population variance | Divide by N−1, not N (Bessel's correction) |
| Sample SD | s = √[Σ(x − x̄)² / (N−1)] |
Sample data; survey results, polls, experiments | s of {2,4,4,4,5,5,7,9} ≈ 2.138 |
| Z-Score | z = (x − μ) / σ |
Standardize a value; compare across distributions | x=7, μ=4, σ=2 → z = (7−4)/2 = 1.5 |
| Coeff. of Variation | CV = (σ / μ) × 100% |
Compare variability of different datasets | σ=5, μ=50 → CV = 10% |
| Standard Error | SE = s / √N |
Precision of sample mean estimate | Smaller SE = more precise estimate of population mean |
| Range | Range = Max − Min |
Simplest spread measure (affected by outliers) | {2,4,6,8,10} → Range = 10 − 2 = 8 |
| IQR | IQR = Q3 − Q1 |
Middle 50% spread; robust to outliers | More reliable than range for skewed data |
How to Calculate Standard Deviation — Step-by-Step
Manual walkthrough of the full calculation process with a real worked example
This is Karl Pearson's classic textbook dataset. Let's calculate the population standard deviation step by step. N = 8 values.
-
1Find the Mean (μ)
Add all values and divide by N.
μ = (2+4+4+4+5+5+7+9) / 8 = 40 / 8 = 5.0 -
2Find Each Deviation from the Mean
Subtract the mean from each data point: (x − μ).
2−5=−3, 4−5=−1, 4−5=−1, 4−5=−1
5−5=0, 5−5=0, 7−5=2, 9−5=4 -
3Square Each Deviation
This eliminates negatives and penalises large deviations.
(−3)²=9, (−1)²=1, (−1)²=1, (−1)²=1
(0)²=0, (0)²=0, (2)²=4, (4)²=16 -
4Sum the Squared Deviations
Add all squared deviations together to get the sum of squares (SS).
SS = 9+1+1+1+0+0+4+16 = 32 -
5Divide by N (for Population) or N−1 (for Sample)
This gives you the variance.
Population Variance: σ² = 32/8 = 4.0
Sample Variance: s² = 32/7 ≈ 4.571 -
6Take the Square Root
The square root of variance gives you the standard deviation — back in original units.
Population SD: σ = √4.0 = 2.0
Sample SD: s = √4.571 ≈ 2.1381
Advertisement
The Normal Distribution & Empirical Rule (68-95-99.7)
How standard deviation relates to the bell curve and what percentages of data fall within each range
In a normal (bell-shaped) distribution, standard deviation has a precise and powerful interpretation. The empirical rule tells us exactly what percentage of data falls within 1, 2, or 3 standard deviations of the mean.
This is why standard deviation is so useful in practice. If you know the mean and standard deviation of a normally distributed dataset, you immediately know the approximate range of most values. For example: if IQ scores have a mean of 100 and SD of 15, then 68% of people have IQ between 85 and 115, and 95% fall between 70 and 130.
What about outliers? Any value beyond ±3σ is considered statistically unusual — it occurs less than 0.3% of the time in a normal distribution. In quality control (Six Sigma manufacturing), targets are set at ±6σ, meaning fewer than 3.4 defects per million.
Real-World Applications of Standard Deviation
Where standard deviation is used across science, medicine, finance, education, and engineering
Population vs Sample Standard Deviation — Deep Dive
When to use σ vs s, why Bessel's correction exists, and common mistakes to avoid
- Divide by N (total count)
- Use when data = entire population
- Examples: heights of all 25 students in a specific class, census of an entire country
- Symbol: σ (Greek sigma)
- Calculator button: σx or σN
- Divide by N−1 (Bessel's correction)
- Use when data = sample from larger group
- Examples: survey of 500 voters from a country of millions, quality checks on 50 items from a factory run
- Symbol: s or SD
- Calculator button: sx or σN-1
In practice: if you're a student or researcher working with experimental data, survey data, or any dataset that doesn't represent every member of the group, use sample SD. In Excel and most calculators, the STDEV() function computes sample SD, while STDEVP() computes population SD.
Does the choice matter? For large N (hundreds or thousands), population and sample SD are nearly identical. The difference becomes significant only for small datasets (N < 30). For N=5, sample SD can be ~12% larger than population SD.
Standard Deviation FAQs
Frequently asked questions about standard deviation, variance, z-scores, and statistics
Sample standard deviation (s) is used when your data is a subset of a larger population. It divides by N−1 (Bessel's correction) to produce an unbiased estimate of the true population standard deviation. When in doubt, use sample SD — it's the safer, more conservative choice.
Dividing by N would systematically underestimate the true population variance (because sample means are closer to their own sample data than to the true population mean). The N−1 denominator inflates the estimate slightly to correct for this bias, making it an unbiased estimator.
Low standard deviation means data points are clustered tightly around the mean — high consistency, precision, or uniformity. Example: a precision machine producing bolts all within ±0.01mm of target.
Whether high or low is "good" depends entirely on context. In manufacturing, you want low SD. In biology studying population diversity, high SD might be expected.
• z = 0: the value equals the mean
• z = +1: one standard deviation above average
• z = −1: one standard deviation below average
• z = +2: unusually high (top ~2.3% in a normal distribution)
• z = −3: extreme outlier (bottom ~0.13%)
In a normal distribution, z-scores let you use standard normal tables to find probabilities. A z-score of 1.96 corresponds to the 97.5th percentile — the basis of the 95% confidence interval.
Variance is expressed in squared units of the original data (e.g., if data is in metres, variance is in metres²), which makes it hard to interpret intuitively. Standard deviation is in the same units as the original data, making it much more interpretable.
Variance is preferred in mathematical derivations and in advanced statistics (ANOVA, regression) because variances of independent variables can be added — standard deviations cannot.
The minimum possible value of SD is 0, which occurs only when all values in the dataset are identical (no variation at all). For example, the dataset {5, 5, 5, 5} has SD = 0.
Use CV when you want to compare variability between datasets with different units or different scales. For example: Dataset A has mean 100 and SD 10 (CV = 10%). Dataset B has mean 500 and SD 40 (CV = 8%). Even though Dataset B has a larger SD, it is relatively less variable than Dataset A.
Note: CV is only meaningful when the mean is positive and not close to zero.
For example: dataset {10, 11, 10, 12, 11} has SD ≈ 0.7. Add one outlier and get {10, 11, 10, 12, 11, 100} — SD jumps to ≈ 34!
When your data contains outliers, consider using median absolute deviation (MAD) or interquartile range (IQR) instead, as these are robust to outliers. Always check for outliers before reporting SD as a summary statistic.