Learning Module 7 Estimation and Inference

50 questions available

Overview and Sampling Methods5 min
This chapter explains how analysts obtain population information through samples and how to quantify sampling error and perform statistical inference. Sampling methods are categorized into probability sampling (simple random, systematic, stratified random, cluster) and non-probability sampling (convenience, judgmental). Simple random sampling gives every population member equal selection probability. Stratified sampling partitions the population into strata and samples proportionally from each stratum to increase precision when subgroups differ. Cluster sampling selects whole clusters and is often cost-effective for large populations but generally less precise. Non-probability sampling is faster and cheaper but risks nonrepresentative samples.
Sampling error is the difference between a statistic and the true population parameter due to using a subset. The sampling distribution of a statistic describes the distribution of that statistic across repeated random samples. The central limit theorem (CLT) states that for a population with finite variance, the sampling distribution of the sample mean is approximately normal with mean equal to the population mean and variance equal to population variance divided by sample size (sigma squared over n) for sufficiently large n. The standard error of the sample mean is sigma/sqrt(n) when population sigma is known, or s/sqrt(n) when estimated from the sample. The CLT justifies use of normal-based confidence intervals and hypothesis tests for large samples; n >= 30 is a common rule of thumb but more may be needed for heavily skewed distributions.

Key Points

  • Probability sampling methods include simple random, stratified, cluster, and systematic sampling.
  • Non-probability sampling includes convenience and judgmental sampling; these can be biased.
  • Sampling error arises because samples are subsets; sampling distribution describes variability of statistics.
  • Central Limit Theorem: sample mean approx. normal with mean mu and variance sigma^2/n for large n.
  • Standard error of sample mean is s / sqrt(n) when sigma unknown.
Resampling: Bootstrap and Jackknife5 min
Resampling methods create empirical sampling distributions when analytic formulas are difficult or assumptions are uncertain. Bootstrap repeatedly draws samples with replacement from the observed sample (each resample same size as original), computes the statistic of interest for each resample, and uses the variation across resamples to estimate standard errors and confidence intervals. The bootstrap estimate of a standard error is the sample standard deviation of the resampled statistics (adjusted for B-1). Jackknife removes one observation at a time (leave-one-out) to estimate bias and standard errors; it uses n replications for a sample of size n and is deterministic for given data.

Key Points

  • Bootstrap draws many resamples with replacement from the observed data to estimate sampling distributions.
  • Bootstrap standard error = sqrt( (1/(B - 1)) sum_{b=1..B} (theta_hat_b - mean_theta_hat)^2 ).
  • Jackknife leaves one observation out at a time to estimate bias and variance; requires n replications.
  • Resampling is valuable when analytic standard errors are hard to obtain or distributional assumptions are weak.
Hypothesis Testing Fundamentals6 min
Hypothesis testing framework: (1) state null (H0) and alternative (Ha) hypotheses (must be mutually exclusive and exhaustive), (2) choose test statistic and its sampling distribution, (3) choose significance level alpha (probability of Type I error), (4) define decision rule (critical values or p-value), (5) compute test statistic from data, (6) decide to reject or fail to reject H0. Type I error (alpha) is rejecting true H0; Type II error (beta) is failing to reject false H0; power = 1 - beta is probability of correctly rejecting false H0. Common test statistics in finance include t for means and regression coefficients, chi-square for variances and contingency tables, F for variance ratios and overall regression fit, and correlation tests using t with df = n - 2.

Key Points

  • Always specify H0 and Ha, sample statistic, distribution, alpha, decision rule, compute and decide.
  • Type I error = alpha (false positive); Type II error = beta (false negative); power = 1 - beta.
  • t-tests are used for means and regression coefficients when sigma unknown (df = n-1 or n-k-1).
  • chi-square test for single variance and contingency-table independence (df = (r-1)(c-1)).
  • F-test for ratio of variances and overall regression significance (MSR/MSE).
Parametric and Nonparametric Tests4 min
Parametric tests rely on distributional assumptions and target parameters (mean, variance). Nonparametric tests (e.g., Wilcoxon signed-rank, Mann-Whitney U, sign test, Spearman rank correlation) make fewer assumptions, work with ranks or ordinal data, are robust to outliers and non-normality, and are useful for small samples or when data are ordinal. However, parametric tests are typically more powerful when their assumptions hold. Guidance is provided on when to use nonparametric alternatives to t-tests and correlation tests.

Key Points

  • Use nonparametric tests when data violate parametric assumptions, contain outliers, or are ordinal.
  • Spearman rank correlation uses ranks and can be tested similarly to Pearson via t for large n.
  • Nonparametric alternatives include Wilcoxon signed-rank for single-sample or paired tests and Mann-Whitney U for two independent samples.
Testing Correlation and Contingency-Table Independence5 min
Testing correlation: Pearson correlation r uses t = r sqrt((n - 2)/(1 - r^2)) with df = n - 2 to test H0: rho = 0. Spearman rank correlation r_s is calculated from rank differences and can be tested via t for large samples. Contingency-table tests: expected count E_{ij} = (row_i_total * col_j_total)/grand_total; chi-square statistic = sum (O_{ij} - E_{ij})^2 / E_{ij}; df = (r - 1)(c - 1). Standardized residuals (Oij - Eij)/sqrt(Eij) indicate which cells differ most from independence.

Key Points

  • Pearson correlation significance uses t with df = n - 2.
  • Spearman correlation is a rank-based nonparametric alternative; for n > 30 use t approximation.
  • Chi-square independence test compares observed vs expected cell frequencies; df = (r-1)(c-1).
  • Standardized residuals help identify which cells drive rejection of independence.

Questions

Question 1

Which sampling method guarantees that every member of a population has an equal chance of selection?

View answer and explanation
Question 2

You want the standard error of a sample mean to be at most 0.5 units. If the population standard deviation is estimated to be 4.0, what minimum sample size n do you need (use the population formula)?

View answer and explanation
Question 3

Which statement best describes the central limit theorem as used for sample means?

View answer and explanation
Question 4

An analyst has a sample of n = 25 independent observations from a population with unknown variance. The sample mean is 10 and the sample standard deviation is 5. What is the estimated standard error of the sample mean?

View answer and explanation
Question 5

Which sampling method is most appropriate when you need to ensure representation across known subgroups (for example, bond duration buckets) and to improve precision?

View answer and explanation
Question 6

An analyst has a sample of 50 returns and computes a 95 percent confidence interval for the mean. Which significance level alpha corresponds to this interval?

View answer and explanation
Question 7

Which statement about bootstrap resampling is true?

View answer and explanation
Question 8

You compute a t-statistic of 2.5 with df = 20 for a two-sided test. Which statement is correct at alpha = 0.05?

View answer and explanation
Question 9

An analyst wants to test whether two independent samples have equal means and assumes equal variances. What test should she use?

View answer and explanation
Question 10

What is the formula for an expected cell frequency Eij in a contingency table under the null hypothesis of independence?

View answer and explanation
Question 11

Which of the following is a valid reason to use a nonparametric test instead of a parametric t-test?

View answer and explanation
Question 12

When performing a bootstrap with B resamples to estimate the standard error of a statistic theta_hat, which expression gives the bootstrap standard error estimate?

View answer and explanation
Question 13

You test H0: sigma^2 = 0.04 for a normally distributed variable using a sample of n = 16 and obtain sample variance s^2 = 0.02. Which test statistic should you use?

View answer and explanation
Question 14

Which of the following best describes a Type I error in hypothesis testing?

View answer and explanation
Question 15

An analyst uses a two-sample t-test with pooled variance and obtains a p-value of 0.03. At alpha = 0.05 what should she conclude?

View answer and explanation
Question 16

Which test is appropriate to examine whether the variance of returns changed after a policy event using two independent samples (before and after) of normally distributed returns?

View answer and explanation
Question 17

A researcher has a single sample of size n = 12 of monthly returns and wishes to estimate the standard error of the sample median but no analytic formula is available. Which method is most appropriate?

View answer and explanation
Question 18

Which of the following is the correct test statistic to test whether a sample Pearson correlation r differs from zero for n observations?

View answer and explanation
Question 19

Which approach is best if you have ranked (ordinal) performance scores for fund managers and you want to test whether two groups have different central tendency?

View answer and explanation
Question 20

You observe two time periods with very different volatilities in a regression residual plot, indicating heteroskedasticity. Which statement is true?

View answer and explanation
Question 21

In a simple linear regression Y = b0 + b1 X + e, what is the meaning of b1?

View answer and explanation
Question 22

Which decomposition relates total variation of Y into explained and unexplained parts in regression?

View answer and explanation
Question 23

In simple linear regression, how is the coefficient of determination R^2 related to the sample correlation r between X and Y?

View answer and explanation
Question 24

You estimated a simple regression with n = 30 observations and obtained SSR = 45 and SSE = 155. What is R^2?

View answer and explanation
Question 25

In regression output, the F-statistic tests which null hypothesis for a simple linear regression?

View answer and explanation
Question 26

Which regression assumption is violated if residuals plotted against time display a clear seasonal pattern?

View answer and explanation
Question 27

You run a simple regression Y on X and obtain slope b1_hat = 0.8, se(b1_hat) = 0.2, n = 25. What is the t-statistic to test H0: b1 = 0?

View answer and explanation
Question 28

Which transformation yields a model where the slope approximates an elasticity (percent change in Y per percent change in X)?

View answer and explanation
Question 29

If residuals of a regression are not normally distributed but the sample size is very large, what does the chapter recommend regarding inference?

View answer and explanation
Question 30

Which of the following is an advantage of cluster sampling relative to simple random sampling?

View answer and explanation
Question 31

An analyst computes sample correlation r = 0.3 with n = 12. Using t = r sqrt((n - 2)/(1 - r^2)), what is the approximate t-statistic (round to two decimals)?

View answer and explanation
Question 32

Which of the following correctly describes the jackknife resampling method?

View answer and explanation
Question 33

If a regression's residuals have a mean not equal to zero, what does the chapter say about that result?

View answer and explanation
Question 34

Which of these is a correct expression for the standard error of the forecast for a new Xf in SLR (prediction standard error)?

View answer and explanation
Question 35

Which functional transformation would you try if plotting Y versus X shows curvature suggesting exponential growth of Y?

View answer and explanation
Question 36

Which of the following is TRUE about the relationship between the t-test for slope and the F-test of overall fit in a simple linear regression?

View answer and explanation
Question 37

Which of the following best describes the purpose of a paired (dependent) samples t-test?

View answer and explanation
Question 38

A contingency table has 4 rows and 3 columns. What are the degrees of freedom for the chi-square test of independence?

View answer and explanation
Question 39

Which statement about the bootstrap and jackknife methods is consistent with the chapter?

View answer and explanation
Question 40

You conduct a chi-square test of independence on a 3x3 contingency table and obtain chi-square statistic = 12.5. The critical value at alpha = 0.05 with df = 4 is 9.49. What is your decision?

View answer and explanation
Question 41

When is the pooled variance estimator used in two-sample t-tests?

View answer and explanation
Question 42

You construct a 95 percent prediction interval for an SLR forecast and find it wide. According to the chapter, which factor would contribute MOST to the interval width?

View answer and explanation
Question 43

Which of the following is an example of non-probability sampling?

View answer and explanation
Question 44

When comparing means for two dependent samples, which distribution does the test statistic follow under normality assumptions?

View answer and explanation
Question 45

An analyst performs stratified sampling with k strata and samples proportional to stratum sizes. Which effect does stratification typically have compared with simple random sampling of the same overall sample size?

View answer and explanation
Question 46

Which of the following best describes the p-value reported by software for a regression coefficient?

View answer and explanation
Question 47

Which of the following is NOT an advantage of bootstrap resampling mentioned in the chapter?

View answer and explanation
Question 48

If you want to test whether two categorical classifications are independent using a sample of 1,000 observations in a 2x3 table, which test and degrees of freedom are appropriate?

View answer and explanation
Question 49

Which diagnostic plot is most helpful to detect heteroskedasticity in a regression model?

View answer and explanation
Question 50

Which of these is a correct interpretation of an R^2 value of 0.80 in a simple linear regression of ROA on CAPEX, as illustrated in the chapter example?

View answer and explanation