This report provides a meticulously curated and exhaustively detailed guide to free online resources for studying the core concepts of statistics as they apply to social and behavioral research in public administration. The document is structured to mirror the provided course syllabus, ensuring direct utility for students and educators. It offers a pedagogical framework that links each statistical tool to its practical application, highlighting not only computational techniques but also the underlying theoretical principles and ethical considerations essential for robust public sector research.
The guide is organized into five units, covering foundational concepts, bivariate relationships, probability theory, statistical estimation, and hypothesis testing. For each topic, it identifies and summarizes high-quality free resources, including Massive Open Online Courses (MOOCs), university lectures, and video tutorials. This synthesis provides a comprehensive roadmap for mastering the tools required to conduct evidence-based research and move from raw data to informed policy decisions.
Introduction: The Foundational Role of Statistics in Public Administration
Statistics is a discipline concerned with the collection, analysis, interpretation, and presentation of data. For students of public administration, this discipline is not merely an academic exercise but a critical toolkit for understanding and improving society. From evaluating the impact of a new public policy on a community’s well-being to predicting population trends for resource allocation, statistical methods provide the quantitative rigor necessary to move beyond anecdote and make decisions grounded in empirical evidence.
This report serves as a navigational aid, guiding the student through the theoretical landscape of statistics as outlined in the course syllabus. It is designed to reinforce the central course objective: to equip future public administrators with the fundamental tools required to conduct social and behavioral research. Each unit builds on the last, progressing from the basic methods of describing data to the complex, yet powerful, techniques of drawing conclusions about large populations from smaller samples. The guide emphasizes that an understanding of these concepts is essential not only for conducting one’s own research but also for critically evaluating the research of others, a core competency in an era of data-driven governance.
Unit 1: Foundational Statistical Concepts
Introduction to Statistics
The field of statistics is fundamentally divided into two major branches: descriptive and inferential statistics.1 Descriptive statistics provides the tools to summarize and organize data in a way that makes it easier to understand. These methods provide a clear picture of a dataset by highlighting key measures and trends. For example, a teacher calculating the average test score for a class is using descriptive statistics to understand overall performance. Similarly, a business analyzing monthly sales data to identify seasonal trends is performing a descriptive analysis. This branch of statistics is focused on what has already been observed, using measures such as the mean, median, mode, and standard deviation. Descriptive statistics can be visually presented through graphs, tables, and charts to help people quickly identify patterns.
Inferential statistics, on the other hand, goes beyond mere description. This branch of statistics uses data from a sample to make broader conclusions, predictions, and inferences about a larger population from which the sample was drawn. Unlike descriptive methods, which only summarize past data, inferential statistics allows a researcher to test hypotheses and make generalizations about an entire population. For instance, a business might use a sample of past sales data to predict future sales, but these predictions come with a margin of error because they are based on a sample rather than the full population. The choice of which branch of statistics to use is guided by the research objective and the type of data analysis required. The process of statistical analysis naturally progresses from descriptive to inferential; one must first be able to summarize their data before being able to draw valid conclusions from it.
Measures of Central Tendency
Measures of central tendency are summary statistics that represent the center point or typical value of a dataset. The three most common measures are the mean, median, and mode.5
- Mean: The mean is the arithmetic average of all values in a dataset. It is calculated by summing all values and dividing by the number of observations.5 This measure is the most commonly used because its calculation incorporates every value in the data.6 The mean is best suited for interval and ratio data, where there is equal spacing between adjacent values, and for data that is symmetrically distributed.5
- Median: The median is the middle value of an ordered dataset.5 For an odd number of observations, the median is the value at the exact middle position. For an even number of observations, the median is the average of the two middle values.5 The median is the preferred measure for datasets with skewed distributions or outliers, as it is not influenced by extreme values.5 For example, U.S. household income data is typically skewed, and the median income provides a more accurate representation of the central tendency than the mean, which can be inflated by a few very high incomes.6
- Mode: The mode is the value that appears most frequently in a dataset.5 Unlike the mean or median, the mode can be used for nominal-level data, where values are categories rather than numbers.5 For example, in a survey of political ideologies, the mode would reveal the most frequently chosen category, such as “Liberal” or “Conservative”.5 A dataset can have one mode, multiple modes, or no mode at all.6
The choice among these measures depends on the data’s level of measurement and distribution.5 While each measure indicates a central point, they do so differently, and using the wrong one can lead to a misleading conclusion about a dataset.
Levels of Measurement | Examples | Measures of Central Tendency |
Nominal | Ethnicity, Political ideology | Mode |
Ordinal | Level of anxiety, Income bracket | Mode, Median |
Interval and ratio | Reaction time, Test score, Temperature | Mode, Median, Mean |
Measures of Dispersion
Measures of central tendency provide only an incomplete picture of a dataset; to fully understand the data, it is crucial to also consider its variability or spread.6 Measures of dispersion quantify how individual data points in a distribution deviate from the central value.
- Variance and Standard Deviation: Variance and standard deviation are measures that show how spread out the data is from the mean.4 Variance is the average of the squared differences from the mean, while standard deviation is the square root of the variance.7 The standard deviation is particularly useful because it is in the same units as the original data, making it easier to interpret.7 A low standard deviation indicates that data points are tightly clustered around the mean, while a high standard deviation suggests that they are more spread out.4
- Coefficient of Variation (CoV): The coefficient of variation is a normalized measure of dispersion that allows for the comparison of variability between different datasets that may have different units or means.8 It is a dimensionless number, expressed as a ratio of the standard deviation to the mean, often as a percentage.8 For a public administrator comparing crime rate data from two cities with different average populations, using the CoV would provide a more accurate comparison of variability than a simple standard deviation, which would be influenced by the scale of each city’s population. It is a powerful tool for understanding data clustering and for direct comparison across different scales.8 However, the CoV is only meaningful for data on a ratio scale with a true zero, such as temperature measured in Kelvin, and can be sensitive to small changes in the mean when the mean value is close to zero.9
Unit 2: Determining the Bi-Variate Relationship
Correlation Analysis
Correlation analysis is used to determine the strength and direction of a linear relationship between two variables.10 The correlation coefficient is a value between -1 and +1 that indicates the degree of this relationship.10 A value of +1 signifies a perfect positive correlation, -1 a perfect negative correlation, and 0 indicates no linear relationship.11
- Karl Pearson’s Correlation Coefficient: This is the most widely used method for measuring correlation.11 It is a measure of the linear relationship between two continuous variables.10 The coefficient is a ratio between the covariance of the two variables and the product of their standard deviations.10 A critical assumption of this method is that a linear relationship exists between the variables.11
- Spearman’s Rank Correlation: Spearman’s method is a non-parametric test that assesses how well the relationship between two variables can be described using a monotonic function.12 This means it is used when the data are ranked or ordinal in nature.12 The calculation involves ranking the data for each variable and then calculating the difference between the ranks.12
It is essential to recognize that correlation does not imply causation.10 A strong correlation between two variables only indicates that they tend to change together, not that one variable’s change causes a change in the other.10 This is a crucial distinction in social research, where confounding variables often influence observed relationships.
Simple Linear Regression
Simple linear regression is a statistical method that allows a researcher to model the relationship between two continuous quantitative variables.15 It moves beyond correlation by fitting a straight line to the data to describe and predict how a dependent variable changes as an independent variable changes.16
The formula for simple linear regression is y=β0+β1x+ϵ.16
- y is the predicted value of the dependent variable.
- β0 is the y-intercept, which is the predicted value of y when x is 0.
- β1 is the regression coefficient, which indicates how much y is expected to change as x increases by one unit.
- x is the independent variable.
- ϵ is the error term, representing the difference between the observed and predicted values.
This method finds the “line of best fit” by minimizing the total error of the model.16 For example, in a study of the relationship between income and happiness, a linear regression model might show that for every additional $10,000 in income, there is a corresponding 0.71-unit increase in reported happiness.16 The model allows for predictions, but a critical limitation is that predictions should not be made outside the range of the data used to create the model.16 The relationship between variables may change at unobserved levels, and extrapolation can lead to invalid conclusions.16
Unit 3: Probability Theory
Basic Terminologies and Approaches to Probability
Probability theory is the mathematical framework for quantifying uncertainty and provides the basis for all statistical inference.17 A probability gives the likelihood that a defined event will occur and is expressed as a number between 0 and 1, where 0 is an impossible event and 1 is a certain one.17 An event is a set of outcomes from a random experiment, and the sample space is the set of all possible outcomes.18
There are three main approaches to assigning probability:
- Classical Approach (Model-based): This approach determines probability based on equally likely outcomes.17 For example, the probability of rolling a “1” on an unbiased six-sided die is 1/6, as there are six equally likely possibilities.17
- Relative Frequency Approach (Frequentist): This approach estimates probability based on observations from a large number of events.17 For instance, if out of 1000 single births, 520 were boys, the probability of a birth being a boy is estimated as 0.52.17
- Subjective Approach: This approach uses personal belief, judgment, or past experience to predict the probability of a future event.17 For example, predicting the probability of a favorite football team winning the next match is a subjective assessment.17
Laws of Probability
To understand the likelihood of multiple events occurring, several fundamental rules of probability are used.
- Addition Rule: This rule is used to find the probability of at least one of two events, A or B, occurring.17
- For mutually exclusive events (they cannot occur at the same time), the probability is the sum of their individual probabilities: P(A or B)=P(A)+P(B).17
- For events that are not mutually exclusive, the probability is the sum of their individual probabilities minus the probability of both occurring simultaneously: P(A or B)=P(A)+P(B)−P(A and B).17
- Multiplication Rule: This rule is used to find the probability that two events, A and B, occur together.17
- For independent events (the occurrence of one does not affect the other), the probability is the product of their individual probabilities: P(A and B)=P(A)×P(B).17
- For events that are not independent, the probability is the product of the probability of the first event and the conditional probability of the second event, given that the first has occurred: P(A and B)=P(A)×P(B∣A).17
- Conditional Probability: This rule defines the probability of an event, A, occurring given that another event, B, has already occurred.17 It is expressed as:
P(A∣B)=P(B)P(A and B).19
These rules provide a systematic framework for quantifying uncertainty and analyzing random phenomena, enabling informed decisions in various fields, from science to finance.20
Rule | Purpose | Mutually Exclusive Events | Non-Mutually Exclusive Events |
Addition Rule | Finds the probability of at least one event occurring. | P(A or B)=P(A)+P(B) | P(A or B)=P(A)+P(B)−P(A and B) |
Multiplication Rule | Finds the probability of both events occurring. | P(A and B)=0 | $P(A \text{ and } B) = P(A) \times P(B |
Theoretical Probability Distributions
Theoretical probability distributions are mathematical functions that describe the probability of different outcomes for a random variable.21 Two key distributions are the binomial and the normal distributions.
- Binomial Distribution: The binomial distribution describes the distribution of discrete data with two possible outcomes, often referred to as “success” or “failure,” from a finite number of independent trials.21 It is defined by two parameters:
n (the number of trials) and p (the probability of success).23 For example, determining whether a person has a specific disease (yes/no) or whether a public policy succeeds or fails would follow a binomial distribution.22 The probability of exactly
x successes in n trials is given by the formula: P(x:n,p)=(xn)px(1−p)n−x.23 - Normal Distribution: The normal distribution describes continuous data that is symmetrically distributed in a characteristic “bell” shape.21 It is completely described by its mean (
μ) and standard deviation (σ).21 This distribution is ubiquitous in statistics because of the
Central Limit Theorem (CLT). The CLT states that if repeated large samples (n≥30) are drawn from any population, regardless of its original shape, the distribution of the sample means will be approximately normal.25 This principle is a cornerstone of inferential statistics, as it justifies the use of a wide range of parametric tests, such as the Z-test, for large samples. It allows researchers to draw robust conclusions about population means even when the underlying population distribution is unknown.26
Unit 4: Statistical Estimation
Concept of Estimation, Sampling Distribution, and Standard Error
Estimation is the process of using data from a sample to make an educated guess about the unknown values of population parameters.25 The process is grounded in the concepts of sampling distributions and standard error.
- Sampling Distribution: A sampling distribution is the probability distribution for a sample statistic, such as the sample mean or sample proportion, based on a fixed-size sample from a population.25 It quantifies the extent to which a statistic will vary from one sample to another due to random chance.25
- Standard Error: The standard error is the standard deviation of a sample statistic’s sampling distribution.25 It is a measure of the uncertainty in a statistical estimate, quantifying the variability of the sample mean around the true population mean.28 This is a crucial distinction from the standard deviation, which measures the variability of individual data points within a single sample.28 The standard error of the mean is a core component of confidence intervals and hypothesis tests.
Criteria of a Good Estimator
A good estimator should possess several key properties to ensure that the estimates it produces are reliable and accurate. These properties include unbiasedness, consistency, efficiency, sufficiency, and robustness.29
- Unbiasedness: An estimator is unbiased if its expected value is equal to the true population parameter.28 This means that, on average, the estimator neither overestimates nor underestimates the parameter.
- Consistency: A consistent estimator is one that converges to the true parameter value as the sample size increases.29 A consistent estimate improves in accuracy as more data is collected.
- Efficiency: An efficient estimator is the one that has the smallest possible margin of error among all unbiased estimators.28 It is the most precise estimator available for the given data, maximizing the information “squeezed” from the sample.29
- Sufficiency: A sufficient estimator captures all the relevant information in the sample about the parameter.29 It does not require any additional data for the estimation.
- Robustness: Robustness refers to an estimator’s resilience to small deviations from its underlying assumptions or to the presence of outliers in the data.29 For example, the median is a more robust estimator of central tendency than the mean when a dataset contains extreme values, such as a billionaire’s income in a neighborhood survey.29
Type of Estimates and Sample Size Determination
In estimation, there are two primary types of estimates:
- Point Estimate: A point estimate is a single number used to estimate a population parameter.25 For example, the sample mean (
xˉ) is a point estimate for the population mean (μ).25 - Interval Estimate: An interval estimate is a range of values, or confidence interval, within which a population parameter is likely to be found.25 It provides a measure of precision around the point estimate and is more informative than a single number.25
Determining the appropriate sample size is a critical step in research design.30 An insufficient sample size can lead to high false negatives and undermine the research’s validity.30 Conversely, an excessively large sample size can be ethically questionable, particularly in studies involving human subjects, and may yield statistically significant but practically meaningless results.30 Key factors to consider for sample size determination include the desired confidence level, the acceptable margin of error, the expected population size, and the study’s power.31 The final decision is often a balance between statistical requirements, available resources, and practical judgment.30
Unit 5: Test of Hypothesis
Introduction to Test of Hypothesis & Core Terminologies
Hypothesis testing is a formal procedure for using sample data to make a decision about a claim or hypothesis concerning a population.34 The process involves formulating two competing hypotheses, collecting data, and using a statistical test to determine whether there is enough evidence to reject a default assumption.34
- Null and Alternative Hypotheses: The process begins with the formulation of two opposing statements.34
- Null Hypothesis (H0): This is the default position, stating that there is no effect, no difference, or no relationship between the variables.34
- Alternative Hypothesis (Ha): This is the hypothesis that the researcher is trying to prove, suggesting that there is a significant effect or relationship.34
- One-tailed vs. Two-tailed Tests: The nature of the alternative hypothesis determines the type of test.36 A one-tailed test is used when the hypothesis specifies a direction (e.g., “greater than” or “less than”).36 A two-tailed test is used when the hypothesis simply states that a parameter is “not equal” to a certain value, without specifying a direction.36
- P-value: The p-value is the probability of observing a test statistic as extreme as, or more extreme than, the one calculated from the sample data, assuming the null hypothesis is true.34 A small p-value (typically less than 0.05) indicates that the observed data would be very unlikely if the null hypothesis were true, thus providing evidence to reject it.34
Type-I and Type-II Errors
In hypothesis testing, a decision is made based on a limited sample, which introduces the possibility of error.35 There are two types of errors that can be made:
- Type I Error (α): This occurs when a true null hypothesis is incorrectly rejected.35 It is also known as a “false positive”.38 The probability of making a Type I error is directly equal to the significance level (
α) chosen for the test.39 - Type II Error (β): This occurs when a false null hypothesis is not rejected.35 This is known as a “false negative”.38
There is an inverse relationship between these two error types.38 Reducing the risk of a Type I error (by lowering
α) increases the risk of a Type II error, and vice versa. For a public administrator, understanding the potential consequences of each error type is crucial. For example, a Type I error might lead to the implementation of an ineffective policy (rejecting the true null that the policy has no effect), while a Type II error might lead to the abandonment of an effective policy (failing to reject the false null that the policy has no effect).35
Test of Significance of Large and Small Samples
Parametric tests are statistical tests that make certain assumptions about the data, often that it follows a specific distribution, such as the normal distribution.29
- Large Sample Tests (Z-tests): For large samples (n≥30), the Central Limit Theorem ensures that the sampling distribution of the mean is approximately normal. In this case, a Z-test is used to determine if a sample mean is significantly different from a population mean.19
- Small Sample Tests (t-tests): When the sample size is small and the population standard deviation is unknown, the t-distribution is used instead of the normal distribution to account for the additional uncertainty.28 A t-test is used to compare the means of one or two small samples.34 The t-distribution is bell-shaped and symmetric, but its shape depends on the degrees of freedom (n-1), making it more appropriate for small samples.25
Parametric and Non-Parametric Tests
The choice between a parametric and a non-parametric test is a critical decision in research design, determined by the properties of the data and the assumptions of the test.13
- Parametric tests assume that the data are drawn from a known distribution and rely on parameters such as the mean and variance.29 Z-tests and t-tests are examples of parametric tests.
- Non-parametric tests make fewer or no assumptions about the data’s distribution and are particularly useful for ordinal or ranked data, or when the data does not meet the assumptions of a parametric test.13
Chi-Square Test
The chi-square (χ2) test is a versatile non-parametric test used for categorical data.34 There are two common types of chi-square tests:
- Goodness-of-Fit Test: This test determines if a single population’s distribution of a categorical variable matches a known or hypothesized distribution.42 For example, it could be used to see if the colors of M&Ms in a bag match the manufacturer’s stated proportions.43
- Test of Independence: This test is used to determine if two categorical variables from a single population are independent or associated.42 The data for this test are organized into a contingency table.44 For example, it could be used to test whether there is a relationship between a person’s physical activity level and their profession.13
Feature | Goodness-of-Fit Test | Test of Independence |
Purpose | To see if a single population fits a known distribution. | To see if two variables are independent. |
Data Structure | A single set of counts. | A contingency table (a matrix of counts). |
Hypotheses | H0: The population fits the given distribution. Ha: The population does not fit the given distribution. | H0: The two variables are independent. Ha: The two variables are dependent. |
Kruskal-Wallis Test
The Kruskal-Wallis H test is a non-parametric alternative to the one-way analysis of variance (ANOVA) and is used to determine if there are statistically significant differences between the mean ranks of two or more independent groups.13 The test is rank-based, meaning it ranks all data points from all groups together and then sums these ranks for each group.13
A key limitation of the Kruskal-Wallis test is that it is an “omnibus” test; it only indicates that at least two groups are different from each other, but it does not specify which groups are different.13 To identify the specific pairs of groups that differ, a post-hoc test, such as the Dunn test, must be performed.41 The Kruskal-Wallis test is particularly useful for social and behavioral research when the data does not meet the assumptions of a parametric test, such as the assumption of a normal distribution.13
Conclusion and Recommendations
The comprehensive nature of the provided syllabus reflects the multifaceted role of statistics in public administration. This report has established that statistics is not a collection of isolated techniques but a coherent framework for understanding and analyzing the world. The concepts are deeply interconnected; for example, the Central Limit Theorem provides the theoretical justification for using Z-tests in hypothesis testing, and a thorough understanding of measures of dispersion is necessary before one can interpret correlation coefficients or regression models.
The curated resources provided for each topic offer a robust and free alternative to traditional academic materials. It is recommended that students approach these resources with a structured learning strategy. They should begin by mastering the foundational concepts of descriptive statistics, as this is the bedrock upon which all subsequent inference and analysis is built. They should use the video tutorials to gain a step-by-step understanding of calculations while referencing the written resources for a deeper theoretical context. Most importantly, as they progress through the course, they should continually link the statistical tools back to their purpose in social and behavioral research, always considering the ethical and practical implications of their analytical choices, from the appropriate selection of a measure to the determination of a sample size. Ultimately, statistics is a tool for revealing truth, and its proper application is a hallmark of responsible public management.