Free Statistics Calculator

Standard Deviation Calculator

Calculate population SD, sample SD, variance, mean, median, z-score, and coefficient of variation — with full step-by-step working shown for every calculation.

Population & Sample SD
Step-by-Step Working
Variance & Z-Score
No Sign-up Required

Standard Deviation Calculator

Enter your dataset — get population SD, sample SD, variance, mean, and more instantly

📝 Test Scores
🌡️ Temperatures
📏 Heights
📈 Stock Returns
🔢 Small Set
Accepts comma-separated, space-separated, or one value per line. Example: 2, 4, 4, 4, 5, 5, 7, 9
Error

Calculation Result

Full statistical breakdown

Standard Deviation
Exact Calculation
Calculation Summary
Step-by-Step Working
Verification — Key Stats Check
All Statistical Measures
Full Details & Interpretation
    Share This Result

    What Is Standard Deviation? A Complete Beginner's Guide

    Understand standard deviation from first principles — definition, notation, formulas, and why it matters across every field

    Standard Deviation: Measuring Spread Around the Mean

    Standard deviation is a measure of how spread out the values in a dataset are relative to the mean (average). A low standard deviation means values are clustered tightly around the mean; a high standard deviation means values are widely scattered. It is the single most widely used measure of statistical dispersion in the world.

    Mathematically, standard deviation is the square root of variance, and variance is the average of the squared differences from the mean. The reason we square the differences is to make all deviations positive (so negatives don't cancel positives) and to penalize large deviations more heavily than small ones.

    📐 Core Insight: Standard deviation answers the question: "On average, how far are the data points from the mean?" If the mean is 50 and the SD is 5, most values fall between 45 and 55. If the SD is 20, values are spread between 30 and 70 — much more variable.

    There are two versions you must know: population standard deviation (σ), used when you have every member of the group, and sample standard deviation (s), used when you have a subset. The difference is whether you divide by N or N−1 — a correction called Bessel's correction that prevents underestimation of the true spread.

    σ
    Population SD (σ)
    Use when your data IS the entire population. Divides by N. Example: you have the heights of every student in one class of 30.
    s
    Sample SD (s)
    Use when your data is a SAMPLE from a larger group. Divides by N−1 (Bessel's correction). Example: you surveyed 30 students from a school of 1,000.
    σ²
    Variance
    The square of SD. Harder to interpret (in squared units) but essential for more advanced statistics like ANOVA, regression analysis, and portfolio theory.
    CV
    Coefficient of Variation (CV)
    SD divided by mean, expressed as a percentage. Allows comparison of variability across datasets with different units or different scales. Low CV = more consistent.

    Standard Deviation Formulas — Complete Reference

    Every formula you need: population, sample, variance, z-score, CV — with worked examples

    The Essential Statistics Formulas
    MeasureFormulaWhen to UseExample
    Mean (Average) μ = Σx / N Central tendency of any dataset Mean of {2,4,6}: (2+4+6)/3 = 4
    Population Variance σ² = Σ(x − μ)² / N Full population data available Squared deviations averaged over all N points
    Population SD σ = √[Σ(x − μ)² / N] Full population; exam grades for entire class σ of {2,4,4,4,5,5,7,9} = 2.0
    Sample Variance s² = Σ(x − x̄)² / (N−1) Sample data; estimate of population variance Divide by N−1, not N (Bessel's correction)
    Sample SD s = √[Σ(x − x̄)² / (N−1)] Sample data; survey results, polls, experiments s of {2,4,4,4,5,5,7,9} ≈ 2.138
    Z-Score z = (x − μ) / σ Standardize a value; compare across distributions x=7, μ=4, σ=2 → z = (7−4)/2 = 1.5
    Coeff. of Variation CV = (σ / μ) × 100% Compare variability of different datasets σ=5, μ=50 → CV = 10%
    Standard Error SE = s / √N Precision of sample mean estimate Smaller SE = more precise estimate of population mean
    Range Range = Max − Min Simplest spread measure (affected by outliers) {2,4,6,8,10} → Range = 10 − 2 = 8
    IQR IQR = Q3 − Q1 Middle 50% spread; robust to outliers More reliable than range for skewed data
    💡 Quick rule: Use population SD when your dataset IS the entire group. Use sample SD when you're drawing conclusions about a larger group from a subset. When in doubt, and especially in research contexts, use sample SD.

    How to Calculate Standard Deviation — Step-by-Step

    Manual walkthrough of the full calculation process with a real worked example

    Calculating SD of: 2, 4, 4, 4, 5, 5, 7, 9

    This is Karl Pearson's classic textbook dataset. Let's calculate the population standard deviation step by step. N = 8 values.

    • 1
      Find the Mean (μ)
      Add all values and divide by N.
      μ = (2+4+4+4+5+5+7+9) / 8 = 40 / 8 = 5.0
    • 2
      Find Each Deviation from the Mean
      Subtract the mean from each data point: (x − μ).
      2−5=−3, 4−5=−1, 4−5=−1, 4−5=−1
      5−5=0, 5−5=0, 7−5=2, 9−5=4
    • 3
      Square Each Deviation
      This eliminates negatives and penalises large deviations.
      (−3)²=9, (−1)²=1, (−1)²=1, (−1)²=1
      (0)²=0, (0)²=0, (2)²=4, (4)²=16
    • 4
      Sum the Squared Deviations
      Add all squared deviations together to get the sum of squares (SS).
      SS = 9+1+1+1+0+0+4+16 = 32
    • 5
      Divide by N (for Population) or N−1 (for Sample)
      This gives you the variance.
      Population Variance: σ² = 32/8 = 4.0
      Sample Variance: s² = 32/7 ≈ 4.571
    • 6
      Take the Square Root
      The square root of variance gives you the standard deviation — back in original units.
      Population SD: σ = √4.0 = 2.0
      Sample SD: s = √4.571 ≈ 2.1381
    Result: The population SD is exactly 2.0 — meaning on average, values are 2 units away from the mean of 5. The dataset {2,4,4,4,5,5,7,9} has moderate spread. You can verify: 5±2 = [3, 7], and indeed most values (2, 4, 4, 4, 5, 5, 7) fall within this range.
    Advertisement

    Advertisement

    The Normal Distribution & Empirical Rule (68-95-99.7)

    How standard deviation relates to the bell curve and what percentages of data fall within each range

    The Empirical Rule: What SD Tells You About Distribution

    In a normal (bell-shaped) distribution, standard deviation has a precise and powerful interpretation. The empirical rule tells us exactly what percentage of data falls within 1, 2, or 3 standard deviations of the mean.

    μ ± 1σ
    68.27%
    of all data falls within one standard deviation of the mean
    μ ± 2σ
    95.45%
    of all data falls within two standard deviations of the mean
    μ ± 3σ
    99.73%
    of all data falls within three standard deviations — almost everything

    This is why standard deviation is so useful in practice. If you know the mean and standard deviation of a normally distributed dataset, you immediately know the approximate range of most values. For example: if IQ scores have a mean of 100 and SD of 15, then 68% of people have IQ between 85 and 115, and 95% fall between 70 and 130.

    What about outliers? Any value beyond ±3σ is considered statistically unusual — it occurs less than 0.3% of the time in a normal distribution. In quality control (Six Sigma manufacturing), targets are set at ±6σ, meaning fewer than 3.4 defects per million.

    📚
    Academic Example: Exam Scores
    Mean = 72, SD = 10. Students in the range 62–82 are "typical" (68%). Students above 92 or below 52 are statistical outliers (<5% of test-takers).
    📈
    Finance Example: Stock Volatility
    Daily return mean = 0.05%, SD = 1.2%. On 95% of days, returns fall between −2.35% and +2.45%. A day outside ±3σ (beyond ±3.65%) is a major market event.

    Real-World Applications of Standard Deviation

    Where standard deviation is used across science, medicine, finance, education, and engineering

    Standard Deviation Powers Every Major Field
    🏥
    Medicine & Healthcare
    Normal ranges for blood pressure, cholesterol, and lab tests are defined as mean ± 2 SD. Patient measurements are flagged as abnormal when they fall outside this range.
    💹
    Finance & Investing
    Volatility of a stock or portfolio is measured by SD of returns. The Sharpe Ratio divides excess return by SD to measure risk-adjusted performance. Options pricing uses SD heavily.
    🎓
    Education & Testing
    Standardized tests (SAT, IQ, GRE) are designed with specific means and SDs. Z-scores convert raw scores to a universal scale. Grading curves use SD to set cut-offs.
    🏭
    Manufacturing & Quality Control
    Six Sigma quality programs aim to keep processes within ±6 SD of target (3.4 defects per million). Control charts use SD to detect when a process drifts out of control.
    🔬
    Science & Research
    Error bars in scientific charts show ±1 SD or ±1 SE. Statistical significance (p-values) relies on SD. Measurement precision and reproducibility are reported in SD terms.
    🤖
    Machine Learning & AI
    Feature normalization (z-score normalization) uses SD. Anomaly detection flags data points beyond a threshold in SD units. Batch normalization in neural networks standardizes by SD.

    Population vs Sample Standard Deviation — Deep Dive

    When to use σ vs s, why Bessel's correction exists, and common mistakes to avoid

    Which Standard Deviation Should You Use?
    Population SD (σ)
    σ = √[Σ(x − μ)² / N]
    • Divide by N (total count)
    • Use when data = entire population
    • Examples: heights of all 25 students in a specific class, census of an entire country
    • Symbol: σ (Greek sigma)
    • Calculator button: σx or σN
    Sample SD (s)
    s = √[Σ(x − x̄)² / (N−1)]
    • Divide by N−1 (Bessel's correction)
    • Use when data = sample from larger group
    • Examples: survey of 500 voters from a country of millions, quality checks on 50 items from a factory run
    • Symbol: s or SD
    • Calculator button: sx or σN-1
    🤔 Why N−1? When you calculate the sample mean (x̄) from the data, you use one "degree of freedom" — meaning the last data point is no longer free to vary. You effectively have N−1 independent pieces of information. Dividing by N would systematically underestimate the population variance; dividing by N−1 corrects this bias. This is called Bessel's correction, named after Friedrich Bessel.

    In practice: if you're a student or researcher working with experimental data, survey data, or any dataset that doesn't represent every member of the group, use sample SD. In Excel and most calculators, the STDEV() function computes sample SD, while STDEVP() computes population SD.

    Does the choice matter? For large N (hundreds or thousands), population and sample SD are nearly identical. The difference becomes significant only for small datasets (N < 30). For N=5, sample SD can be ~12% larger than population SD.

    Standard Deviation FAQs

    Frequently asked questions about standard deviation, variance, z-scores, and statistics

    What is the difference between population and sample standard deviation?
    Population standard deviation (σ) is used when your data represents every single member of the group you are studying. It divides the sum of squared deviations by N.

    Sample standard deviation (s) is used when your data is a subset of a larger population. It divides by N−1 (Bessel's correction) to produce an unbiased estimate of the true population standard deviation. When in doubt, use sample SD — it's the safer, more conservative choice.
    Why does sample standard deviation divide by N−1?
    Dividing by N−1 instead of N is called Bessel's correction. When you estimate the mean from the same sample data, you use one degree of freedom, leaving only N−1 independent pieces of information.

    Dividing by N would systematically underestimate the true population variance (because sample means are closer to their own sample data than to the true population mean). The N−1 denominator inflates the estimate slightly to correct for this bias, making it an unbiased estimator.
    What does a high or low standard deviation mean?
    High standard deviation means data points are spread widely from the mean — high variability, inconsistency, or diversity in the dataset. Example: exam scores from 10 to 100 with a mean of 55 and SD of 25.

    Low standard deviation means data points are clustered tightly around the mean — high consistency, precision, or uniformity. Example: a precision machine producing bolts all within ±0.01mm of target.

    Whether high or low is "good" depends entirely on context. In manufacturing, you want low SD. In biology studying population diversity, high SD might be expected.
    How do you interpret a z-score?
    A z-score (or standard score) tells you how many standard deviations a specific value is from the mean. Formula: z = (x − μ) / σ.

    • z = 0: the value equals the mean
    • z = +1: one standard deviation above average
    • z = −1: one standard deviation below average
    • z = +2: unusually high (top ~2.3% in a normal distribution)
    • z = −3: extreme outlier (bottom ~0.13%)

    In a normal distribution, z-scores let you use standard normal tables to find probabilities. A z-score of 1.96 corresponds to the 97.5th percentile — the basis of the 95% confidence interval.
    What is variance and how does it relate to standard deviation?
    Variance is the average of the squared deviations from the mean. Standard deviation is simply the square root of variance.

    Variance is expressed in squared units of the original data (e.g., if data is in metres, variance is in metres²), which makes it hard to interpret intuitively. Standard deviation is in the same units as the original data, making it much more interpretable.

    Variance is preferred in mathematical derivations and in advanced statistics (ANOVA, regression) because variances of independent variables can be added — standard deviations cannot.
    Can standard deviation be negative?
    No, standard deviation can never be negative. Here's why: it involves squaring all deviations (making them ≥ 0), summing them (still ≥ 0), dividing by N or N−1 (still ≥ 0), then taking the square root (the positive root).

    The minimum possible value of SD is 0, which occurs only when all values in the dataset are identical (no variation at all). For example, the dataset {5, 5, 5, 5} has SD = 0.
    What is the coefficient of variation (CV) and when should I use it?
    The Coefficient of Variation (CV) is the ratio of the standard deviation to the mean, usually expressed as a percentage: CV = (σ / μ) × 100%.

    Use CV when you want to compare variability between datasets with different units or different scales. For example: Dataset A has mean 100 and SD 10 (CV = 10%). Dataset B has mean 500 and SD 40 (CV = 8%). Even though Dataset B has a larger SD, it is relatively less variable than Dataset A.

    Note: CV is only meaningful when the mean is positive and not close to zero.
    How do outliers affect standard deviation?
    Standard deviation is highly sensitive to outliers because deviations are squared — so a value far from the mean contributes disproportionately to the sum of squares.

    For example: dataset {10, 11, 10, 12, 11} has SD ≈ 0.7. Add one outlier and get {10, 11, 10, 12, 11, 100} — SD jumps to ≈ 34!

    When your data contains outliers, consider using median absolute deviation (MAD) or interquartile range (IQR) instead, as these are robust to outliers. Always check for outliers before reporting SD as a summary statistic.