• Descriptive
    • Moments
    • Concentration
    • Central Tendency
    • Variability
    • Stem-and-Leaf Plot
    • Histogram & Frequency Table
    • Data Quality Forensics
    • Conditional EDA
    • Quantiles
    • Kernel Density Estimation
    • Normal QQ Plot
    • Bootstrap Plot

    • Multivariate Descriptive Statistics
  • Distributions
    • Binomial Probabilities
    • Geometric Probabilities
    • Negative Binomial Probabilities
    • Hypergeometric Probabilities
    • Multinomial Probabilities
    • Poisson Probabilities

    • Exponential
    • Gamma
    • Erlang
    • Weibull
    • Rayleigh
    • Lognormal
    • Pareto
    • Inverse Gamma

    • Beta
    • Power
    • Beta Prime (Inv. Beta)
    • Triangular

    • Normal (area)
    • Logistic
    • Laplace
    • Cauchy (standard)
    • Cauchy (location-scale)
    • Gumbel

    • Normal RNG
    • ML Fitting
    • Tukey Lambda PPCC
    • Box-Cox Normality Plot
    • Sample Correlation r

    • Empirical Tests
  • Hypotheses
    • Theoretical Aspects of Hypothesis Testing
    • Bayesian Inference
    • Minimum Sample Size

    • Empirical Tests
    • Multivariate (pair-wise) Testing
  • Models
    • Manual Model Building
  • Time Series
    • Time Series Plot
    • Decomposition
    • Exponential Smoothing

    • Blocked Bootstrap Plot
    • Mean Plot
    • (P)ACF
    • VRM
    • Standard Deviation-Mean Plot
    • Spectral Analysis
    • ARIMA

    • Cross Correlation Function
    • Granger Causality
  1. Hypothesis Testing
  2. 101  Statistical Test of the difference between Means -- Dependent/Paired Samples
  • Preface
  • Getting Started
    • 1  Introduction
    • 2  Why Do We Need Innovative Technology?
    • 3  Basic Definitions
    • 4  The Big Picture: Why We Analyze Data
  • Introduction to Probability
    • 5  Definitions of Probability
    • 6  Jeffreys’ axiom system
    • 7  Bayes’ Theorem
    • 8  Sensitivity and Specificity
    • 9  Naive Bayes Classifier
    • 10  Law of Large Numbers

    • 11  Problems
  • Probability Distributions
    • 12  Bernoulli Distribution
    • 13  Binomial Distribution
    • 14  Geometric Distribution
    • 15  Negative Binomial Distribution
    • 16  Hypergeometric Distribution
    • 17  Multinomial Distribution
    • 18  Poisson Distribution

    • 19  Uniform Distribution (Rectangular Distribution)
    • 20  Normal Distribution (Gaussian Distribution)
    • 21  Gaussian Naive Bayes Classifier
    • 22  Chi Distribution
    • 23  Chi-squared Distribution (1 parameter)
    • 24  Chi-squared Distribution (2 parameters)
    • 25  Student t-Distribution
    • 26  Fisher F-Distribution
    • 27  Exponential Distribution
    • 28  Lognormal Distribution
    • 29  Gamma Distribution
    • 30  Beta Distribution
    • 31  Weibull Distribution
    • 32  Pareto Distribution
    • 33  Inverse Gamma Distribution
    • 34  Rayleigh Distribution
    • 35  Erlang Distribution
    • 36  Logistic Distribution
    • 37  Laplace Distribution
    • 38  Gumbel Distribution
    • 39  Cauchy Distribution
    • 40  Triangular Distribution
    • 41  Power Distribution
    • 42  Beta Prime Distribution
    • 43  Sample Correlation Distribution

    • 44  Problems
  • Descriptive Statistics & Exploratory Data Analysis
    • 45  Types of Data
    • 46  Datasheets

    • 47  Frequency Plot (Bar Plot)
    • 48  Frequency Table
    • 49  Contingency Table
    • 50  Binomial Classification Metrics
    • 51  Confusion Matrix
    • 52  ROC Analysis

    • 53  Stem-and-Leaf Plot
    • 54  Histogram
    • 55  Data Quality Forensics
    • 56  Quantiles
    • 57  Central Tendency
    • 58  Variability
    • 59  Skewness & Kurtosis
    • 60  Concentration
    • 61  Notched Boxplot
    • 62  Scatterplot
    • 63  Pearson Correlation
    • 64  Rank Correlation
    • 65  Partial Pearson Correlation
    • 66  Simple Linear Regression
    • 67  Moments
    • 68  Quantile-Quantile Plot (QQ Plot)
    • 69  Normal Probability Plot
    • 70  Probability Plot Correlation Coefficient Plot (PPCC Plot)
    • 71  Box-Cox Normality Plot
    • 72  Kernel Density Estimation
    • 73  Bivariate Kernel Density Plot
    • 74  Conditional EDA: Panel Diagnostics
    • 75  Bootstrap Plot (Central Tendency)
    • 76  Survey Scores Rank Order Comparison
    • 77  Cronbach Alpha

    • 78  Equi-distant Time Series
    • 79  Time Series Plot (Run Sequence Plot)
    • 80  Mean Plot
    • 81  Blocked Bootstrap Plot (Central Tendency)
    • 82  Standard Deviation-Mean Plot
    • 83  Variance Reduction Matrix
    • 84  (Partial) Autocorrelation Function
    • 85  Periodogram & Cumulative Periodogram

    • 86  Problems
  • Hypothesis Testing
    • 87  Normal Distributions revisited
    • 88  The Population
    • 89  The Sample
    • 90  The One-Sided Hypothesis Test
    • 91  The Two-Sided Hypothesis Test
    • 92  When to use a one-sided or two-sided test?
    • 93  What if \(\sigma\) is unknown?
    • 94  The Central Limit Theorem (revisited)
    • 95  Statistical Test of the Population Mean with known Variance
    • 96  Statistical Test of the Population Mean with unknown Variance
    • 97  Statistical Test of the Variance
    • 98  Statistical Test of the Population Proportion
    • 99  Statistical Test of the Standard Deviation \(\sigma\)
    • 100  Statistical Test of the difference between Means -- Independent/Unpaired Samples
    • 101  Statistical Test of the difference between Means -- Dependent/Paired Samples
    • 102  Statistical Test of the difference between Variances -- Independent/Unpaired Samples

    • 103  Hypothesis Testing for Research Purposes
    • 104  Decision Thresholds, Alpha, and Confidence Levels
    • 105  Bayesian Inference for Decision-Making
    • 106  One Sample t-Test
    • 107  Skewness & Kurtosis Tests
    • 108  Paired Two Sample t-Test
    • 109  Wilcoxon Signed-Rank Test
    • 110  Unpaired Two Sample t-Test
    • 111  Unpaired Two Sample Welch Test
    • 112  Two One-Sided Tests (TOST) for Equivalence
    • 113  Mann-Whitney U test (Wilcoxon Rank-Sum Test)
    • 114  Bayesian Two Sample Test
    • 115  Median Test based on Notched Boxplots
    • 116  Chi-Squared Tests for Count Data
    • 117  Kolmogorov-Smirnov Test
    • 118  One Way Analysis of Variance (1-way ANOVA)
    • 119  Kruskal-Wallis Test
    • 120  Two Way Analysis of Variance (2-way ANOVA)
    • 121  Repeated Measures ANOVA
    • 122  Friedman Test
    • 123  Testing Correlations
    • 124  A Note on Causality

    • 125  Problems
  • Regression Models
    • 126  Simple Linear Regression Model (SLRM)
    • 127  Multiple Linear Regression Model (MLRM)
    • 128  Logistic Regression
    • 129  Generalized Linear Models
    • 130  Multinomial and Ordinal Logistic Regression
    • 131  Cox Proportional Hazards Regression
    • 132  Conditional Inference Trees
    • 133  Leaf Diagnostics for Conditional Inference Trees
    • 134  Hypothesis Testing with Linear Regression Models (from a Practical Point of View)

    • 135  Problems
  • Introduction to Time Series Analysis
    • 136  Case: the Market of Health and Personal Care Products
    • 137  Decomposition of Time Series
    • 138  Ad hoc Forecasting of Time Series
  • Box-Jenkins Analysis
    • 139  Introduction to Box-Jenkins Analysis
    • 140  Theoretical Concepts
    • 141  Stationarity
    • 142  Identifying ARMA parameters
    • 143  Estimating ARMA Parameters and Residual Diagnostics
    • 144  Forecasting with ARIMA models
    • 145  Intervention Analysis
    • 146  Cross-Correlation Function
    • 147  Transfer Function Noise Models
    • 148  General-to-Specific Modeling
  • References
  • Appendices
    • Appendices
    • A  Method Selection Guide
    • B  Presentations and Teaching Materials
    • C  R Language Concepts for Statistical Computing
    • D  Matrix Algebra
    • E  Standard Normal Table (Gaussian Table)
    • F  Critical values of Student’s \(t\) distribution with \(\nu\) degrees of freedom
    • G  Upper-tail critical values of the \(\chi^2\)-distribution with \(\nu\) degrees of freedom
    • H  Lower-tail critical values of the \(\chi^2\)-distribution with \(\nu\) degrees of freedom

Table of contents

  • 101.1 Theory
  • 101.2 Example
    • 101.2.1 Critical Value (Region)
    • 101.2.2 P-Value
DRAFT This draft is under development — DO NOT CITE OR SHARE.
  1. Hypothesis Testing
  2. 101  Statistical Test of the difference between Means -- Dependent/Paired Samples

101  Statistical Test of the difference between Means -- Dependent/Paired Samples

101.1 Theory

We define a first population \(X_1 \sim \text{N}\left( \mu_1, \sigma_1^2 \right)\) from which a simple random sample is drawn of size \(n\) with sample mean \(\bar{x}_1 = \frac{1}{n} \sum_{i=1}^{n}x_{1i}\).

We also define a second population \(X_2 \sim \text{N}\left( \mu_2, \sigma_2^2 \right)\) from which a simple random sample is drawn of size \(n\) with sample mean \(\bar{x}_2 = \frac{1}{n} \sum_{i=1}^{n}x_{2i}\).

To test the difference between the sample means we compute the following test statistic

\[ \begin{align*}\bar{d} &= \bar{x}_1 - \bar{x}_2 \\&= \frac{1}{n} \sum_{i=1}^{n} \left( x_{1i} - x_{2i} \right) \\&= \frac{1}{n} \sum_{i=1}^{n} d_i\end{align*} \]

which has the following distribution

\[ \bar{d} \sim \text{N} \left( D, \frac{\sigma_D^2}{n} \right) \]

where \(D = \mu_1 - \mu_2\).

Now it is possible to transform the test statistic in a convenient form

\[ u = \frac{\bar{d} - D}{\frac{\sigma_D}{\sqrt{n}}} \sim \text{N}(0,1) \]

which can be used with the Standard Normal Table of Appendix E.

If the population variance of the difference is unknown this becomes

\[ t = \frac{\bar{d} - D}{\frac{s_d}{\sqrt{n}}} \sim t_{n-1} \]

with

\[ s_d^2 = \frac{1}{n-1} \sum_{i=1}^{n} \left[ \left( x_{1i} - x_{2i} \right) - \bar{d} \right]^2 = \frac{1}{n-1} \sum_{i=1}^{n} \left( d_i - \bar{d} \right)^2 \]

101.2 Example

Consider the following paired samples

Observation Sample 1 Sample 2 \(d_i\) \(d_i^2\)
1 106 102 4 16
2 98 94 4 16
3 123 118 5 25
4 97 91 6 36
5 88 83 5 25
118

Now we can compute

\[ \bar{d} = \frac{24}{5} = 4.8 \text{ or } \bar{d} = \bar{x}_1 - \bar{x}_2 = 102.4 - 97.6 = 4.8 \]

and, based on the alternative formulation of the Variance (Section 58.5.2), we compute

\[ \tilde{s}_d^2 = \frac{1}{n} \sum_{i=1}^{n} \left( d_i - \bar{d} \right)^2 = \frac{1}{n} \sum_{i=1}^{n} d_i^2 - \bar{d}^2 = \frac{118}{5} - (4.8)^2 = 0.56 \]

An unbiased estimate for the variance of the differences is given by

\[ s_d^2 = \frac{1}{n-1} \sum_{i=1}^{n} \left( d_i - \bar{d} \right)^2 = \tilde{s}_d^2 \times \frac{n}{n-1} = 0.56 \times \frac{5}{5-1} = 0.7 \]

which results in

\[ s_d = \sqrt{0.7} = 0.8367 \]

101.2.1 Critical Value (Region)

Because the alternative is right-sided (\(H_A: D > D_0 = 0\)), the critical region is of the form \(\bar{d} \geq c\).

\[ \text{P} (\bar{d} \geq c) = 0.05 \]

\[ \text{P} \left( \frac{\bar{d} - D}{\frac{s_d}{\sqrt{n}}} \geq \frac{c - D}{\frac{s_d}{\sqrt{n}}} \right) = 0.05 \]

If we define the following hypotheses

\[ \begin{cases}\text{H}_0: D_0 = \mu_1 - \mu_2 = 0 \\\text{H}_A: D > D_0 = 0\end{cases} \]

Then we can write

\[ \text{P} \left( t \geq \frac{c}{\frac{s_d}{\sqrt{n}}} \right) = 0.05 \]

from which it follows that

\[ \begin{align*}\frac{c \sqrt{n}}{s_d} &= 2.132 \\c &= 2.132 \times \frac{s_d}{\sqrt{n}} \\&= 2.132 \times \frac{0.8367}{2.236} \\&= 0.7978\end{align*} \]

We conclude that

\[ \text{P}\left( \bar{d} \geq 0.7978 \right) = 0.05 \]

Since \(\bar{d} = 4.8\) is larger than the critical value \(c=0.7978\) we reject the Null Hypothesis H\(_0: D_0 = \mu_1 - \mu_2 = 0\) and accept the Alternative Hypothesis.

101.2.2 P-Value

\[ t = \frac{\bar{d}}{\frac{s_d}{\sqrt{n}}} = \frac{\bar{d}\sqrt{n}}{s_d} = \frac{4.8 \times 2.236}{0.8367} = 12.828 \]

Since the number of degrees of freedom df = \(5-1=4\) this can be written as

\[ \text{P}(t_{4} \geq 12.828) = 0.0001 \]

Since the probability 0.0001 is smaller than \(\alpha = 0.05\) we reject the Null Hypothesis H\(_0: D_0 = \mu_1 - \mu_2 = 0\) and accept the Alternative Hypothesis.

100  Statistical Test of the difference between Means -- Independent/Unpaired Samples
102  Statistical Test of the difference between Variances -- Independent/Unpaired Samples

© 2026 Patrick Wessa. Provided as-is, without warranty.

Feedback: e-mail | Anonymous contributions: click to copy (Sats) | click to copy (XMR)

Cookie Preferences