1. Descriptive Statistics

Descriptive statistics summarize and describe the main features of a dataset. They form the foundation for all statistical analysis and inference.

Measures of Central Tendency

Arithmetic Mean
x̄ = (1/n) ∑i=1n xi
Median
Median = x((n+1)/2) for odd n
Median = (x(n/2) + x(n/2+1)) / 2 for even n

Mode: The most frequently occurring value in the dataset.

Measures of Dispersion

Sample Variance
s2 = (1/(n-1)) ∑i=1n (xi - x̄)2
The n-1 divisor (Bessel's correction) provides an unbiased estimate of population variance.
Standard Deviation
s = √s2
Coefficient of Variation
CV = (s / x̄) × 100%
Dimensionless measure useful for comparing variability across different scales.

Measures of Shape

Skewness
g1 = (1/n) ∑ [(xi - x̄) / s]3
Positive skewness: tail extends to the right. Negative skewness: tail extends to the left.
Kurtosis
g2 = (1/n) ∑ [(xi - x̄) / s]4 - 3
Excess kurtosis (subtracting 3) measures deviation from normal distribution. Positive = heavier tails.

2. Probability Theory

Axioms of Probability (Kolmogorov)

  1. P(A) ≥ 0 for any event A
  2. P(S) = 1 where S is the sample space
  3. For mutually exclusive events: P(A ∪ B) = P(A) + P(B)

Conditional Probability

P(A|B) = P(A ∩ B) / P(B), provided P(B) > 0

Bayes' Theorem

Bayes' Theorem
P(A|B) = P(B|A) P(A) / P(B)
Relates conditional probabilities and allows updating beliefs based on new evidence.

Law of Total Probability

P(B) = ∑i P(B|Ai) P(Ai)
Where {Ai} forms a partition of the sample space.

3. Probability Distributions

Discrete Distributions

Binomial Distribution

P(X = k) = C(n,k) pk (1-p)n-k
E[X] = np, Var(X) = np(1-p)
Models the number of successes in n independent Bernoulli trials.

Poisson Distribution

P(X = k) = e λk / k!
E[X] = Var(X) = λ
Models rare events over time or space.

Continuous Distributions

Normal Distribution

Probability Density Function
f(x) = (1 / σ√(2π)) exp[-(x - μ)2 / (2σ2)]
The most important distribution due to the Central Limit Theorem.

Standard Normal Distribution

Z = (X - μ) / σ ~ N(0, 1)

t-Distribution

Used when population variance is unknown and estimated from sample. Has heavier tails than normal; approaches normal as df → ∞.

Chi-Square Distribution

If Z1, ..., Zk are independent standard normal, then ∑Zi2 ~ χ2(k)

F-Distribution

Ratio of two independent chi-square variables divided by their degrees of freedom. Used in ANOVA and regression.

4. Point Estimation

Properties of Estimators

Maximum Likelihood Estimation

Likelihood Function
L(θ) = ∏i=1n f(xi; θ)
MLE finds θ̂ that maximizes L(θ), typically by setting ∂log L / ∂θ = 0

5. Confidence Intervals

A (1-α) confidence interval provides a range that, in repeated sampling, would contain the true parameter value (1-α)×100% of the time.

CI for Mean (σ known)
x̄ ± zα/2 (σ / √n)
CI for Mean (σ unknown)
x̄ ± tα/2, n-1 (s / √n)
CI for Proportion
p̂ ± zα/2 √(p̂(1-p̂) / n)

6. Hypothesis Testing Framework

The Testing Process

  1. State null (H0) and alternative (H1) hypotheses
  2. Choose significance level α
  3. Select appropriate test statistic
  4. Determine critical region or compute p-value
  5. Make decision: reject or fail to reject H0

Types of Errors

H0 True H0 False
Reject H0Type I Error (α)Correct Decision (Power)
Fail to RejectCorrect DecisionType II Error (β)
Statistical Power
Power = 1 - β = P(Reject H0 | H0 is false)

7. t-Tests

One-Sample t-Test

t = (x̄ - μ0) / (s / √n)
Tests whether the sample mean differs from a hypothesized value μ0. df = n - 1

Independent Two-Sample t-Test

t = (x̄1 - x̄2) / √(sp2(1/n1 + 1/n2))
Where sp2 is the pooled variance (assuming equal variances).

Paired t-Test

t = d̄ / (sd / √n)
Where d̄ is the mean of paired differences. Used for before-after designs or matched pairs.

8. Analysis of Variance (ANOVA)

One-Way ANOVA

Tests whether means of k groups are equal: H0: μ1 = μ2 = ... = μk

F-Statistic
F = MSBetween / MSWithin = (SSB/(k-1)) / (SSW/(N-k))
Sum of Squares
SSTotal = SSBetween + SSWithin

SSB = ∑j nj(x̄j - x̄)2

SSW = ∑ji (xij - x̄j)2

Assumptions of ANOVA

Post-Hoc Tests

When ANOVA rejects H0, post-hoc tests identify which means differ:

9. Nonparametric Tests

Nonparametric tests make fewer assumptions about the underlying distribution and are appropriate when normality cannot be assumed or with ordinal data.

Mann-Whitney U Test

Nonparametric alternative to independent two-sample t-test. Tests whether one distribution is stochastically greater than the other.

Wilcoxon Signed-Rank Test

Nonparametric alternative to paired t-test. Uses ranks of absolute differences.

Kruskal-Wallis Test

Nonparametric alternative to one-way ANOVA. Extends Mann-Whitney to k groups.

10. Correlation Analysis

Pearson Correlation Coefficient
r = ∑(xi - x̄)(yi - ȳ) / √[∑(xi - x̄)2 ∑(yi - ȳ)2]
Measures linear association. -1 ≤ r ≤ 1

Testing Correlation

t = r √(n-2) / √(1-r2)
Tests H0: ρ = 0. df = n - 2

Spearman Rank Correlation

Nonparametric correlation based on ranks. Measures monotonic (not necessarily linear) relationships.

11. Regression Analysis

Simple Linear Regression

Model
Y = β0 + β1X + ε
Least Squares Estimates
b1 = ∑(xi - x̄)(yi - ȳ) / ∑(xi - x̄)2

b0 = ȳ - b1

Coefficient of Determination

R2 = 1 - SSResidual / SSTotal = SSRegression / SSTotal
Proportion of variance in Y explained by the model. 0 ≤ R2 ≤ 1

Regression Inference

t-Test for Slope
t = b1 / SE(b1)
Tests H0: β1 = 0 (no linear relationship)

12. Categorical Data Analysis

Chi-Square Test of Independence

χ2 = ∑ (Oij - Eij)2 / Eij
Where Eij = (row total)(column total) / grand total
df = (r-1)(c-1)

Chi-Square Goodness of Fit

Tests whether observed frequencies match expected frequencies from a hypothesized distribution.

Fisher's Exact Test

Exact test for 2×2 tables when expected frequencies are small (any Eij < 5).

References and Further Reading

  1. Casella, G. & Berger, R.L. (2002). Statistical Inference, 2nd Edition. Cengage Learning.
  2. Wackerly, D.D., Mendenhall, W., & Scheaffer, R.L. (2014). Mathematical Statistics with Applications, 7th Edition. Cengage.
  3. Agresti, A. (2018). Statistical Methods for the Social Sciences, 5th Edition. Pearson.
  4. Kutner, M.H., et al. (2004). Applied Linear Statistical Models, 5th Edition. McGraw-Hill.
  5. Hollander, M., Wolfe, D.A., & Chicken, E. (2013). Nonparametric Statistical Methods, 3rd Edition. Wiley.