QE Comprehensive Revision Sheet (Accessible)

SCREEN READER ADVISORY ON CASE SENSITIVITY: In this document, the case of letters is vital:

1. Mathematical & Statistical Foundations

Expected Value (Expectation)

The "average" value of a random variable in the population.

Linearity: E[aX + bY] = a E[X] + b E[Y]

Variance and Standard Deviation

Measures the spread or "risk" of a variable.

Var(X) = E[X^2] - (E[X])^2

Properties: Var(aX + b) = a^2 * Var(X).

Covariance and Correlation

Measures the linear relationship between two variables.

Cov(X, Y) = E[XY] - E[X]E[Y]
Corr(X, Y) = Cov(X, Y) / [ SD(X) * SD(Y) ]

Conditional Expectation & LIE

Law of Iterated Expectations (LIE): E[Y] = E[ E[Y|X] ]

Independence: If X and Y are independent, then E[Y|X] = E[Y] and Cov(X, Y) = 0.

Indicator Function (The Math behind Dummies)

An indicator function, often written as 1(.) or I(.), translates a logical condition into a number.

1(A) = { 1 if A occurs; 0 if A does not occur }

In Econometrics: This is used to create dummy variables. For example, if we define D = 1(Female), then D=1 for female observations and D=0 for male observations. The expectation of an indicator function is the probability of the event: E[1(A)] = P(A).

2. The Linear Regression Model

The Population Model

Y_i = beta_0 + beta_1*X_1i + beta_2*X_2i + ... + u_i

The Sample Model (Predictions)

Y_i_hat = beta_0_hat + beta_1_hat * X_1i + ...

Dummy Variables

Binary variables taking values 0 or 1. Used for categories (e.g., gender, region).

The Dummy Variable Trap: If you have G categories, you must only include G-1 dummy variables if you have an intercept (constant) in the model. Otherwise, you face perfect multicollinearity.

3. OLS Estimation & Derivation

The OLS Criterion

We choose the beta-hats to minimize the Sum of Squared Residuals (SSR).

Minimize: Sum [ (Y_i - beta_0 - beta_1*X_i)^2 ]

Simple OLS Estimators

beta_1_hat = Sample_Cov(X, Y) / Sample_Var(X)
beta_0_hat = Y_bar - (beta_1_hat * X_bar)

Goodness of Fit

4. The OLS Assumptions (Gauss-Markov)

To ensure OLS is the Best Linear Unbiased Estimator (BLUE):

  1. Linearity in Parameters: The model is a linear combination of betas.
  2. Random Sampling: (X_i, Y_i) are i.i.d.
  3. No Perfect Multicollinearity: Regressors aren't perfectly redundant.
  4. Zero Conditional Mean (Orthogonality): E[u|X] = 0. This is the most critical assumption!
  5. Homoskedasticity: Var(u|X) = sigma^2 (constant). If violated, we have Heteroskedasticity and need "Robust Standard Errors."

5. Statistical Inference

Hypothesis Testing Recipe

  1. Null Hypothesis (H0): Usually beta_1 = 0 (no effect).
  2. Alternative (H1): beta_1 is not 0.
  3. Calculate t-statistic:
    t = (beta_hat - hyper_null) / SE(beta_hat)
  4. Decision Rule: If |t| > 1.96, reject H0 at 5% significance level.

p-values and Confidence Intervals

The F-Statistic

Used to test multiple restrictions simultaneously (e.g., test if beta_1 = 0 AND beta_2 = 0).

6. Large Sample Theory (Asymptotics)

Law of Large Numbers (LLN)

Sample averages converge to population averages as n becomes large.

Y_bar ->p mu_Y

Assures that OLS is Consistent (beta_hat ->p beta).

Central Limit Theorem (CLT)

The CLT is the reason we can perform hypothesis tests (like t-tests) without knowing the underlying distribution of the errors, provided the sample size n is large.

1. The Formal Definition

As the sample size n increases, the distribution of the standardized sample mean converges to a Standard Normal distribution N(0, 1).

sqrt(n) * ( (Y_bar - mu) / sigma ) →d N(0, 1)

2. The "Root-n" Logic (√n)

Why do we multiply by √n? As n gets larger:

3. Application to OLS (The Derivation)

To see why OLS estimators are Asymptotically Normal, we decompose the slope estimator:

beta_1_hat - beta_1 = [ Sum( (X_i - X_bar) * u_i ) ] / [ Sum( (X_i - X_bar)^2 ) ]

Step-by-Step Logic:

  1. Numerator: The term Sum( (X_i - X_bar) * u_i ) is a sum of independent random variables. By the CLT, this sum (when scaled by 1/√n) becomes Normally distributed.
  2. Denominator: By the Law of Large Numbers (LLN), the average (1/n) * Sum( (X_i - X_bar)^2 ) converges to the population variance Var(X).
  3. Conclusion: Since the numerator is Normal and the denominator is a constant (in large samples), their ratio is also Normal.
Bottom Line: In small samples, β_hat is only normal if the errors u are normal. In large samples, β_hat is normal even if the errors are not normal. This allows us to use the 1.96 critical value for almost any large-sample regression.

7. Nonlinear Specifications

Logarithms (Percentages)

Polynomials (Curvatures)

Model: Y = beta_0 + beta_1*X + beta_2*X^2 + u.

To find the effect of X on Y, you must take the derivative: dY/dX = beta_1 + 2*beta_2*X.

Interactions

Model: Y = beta_0 + beta_1*X + beta_2*D + beta_3*(X*D) + u.

The effect of X on Y now depends on whether D is 0 or 1.

8. Internal Validity & Sources of Bias

Internal validity asks: "Is our estimate of the causal effect correct for our sample?"

1. Omitted Variable Bias (OVB)

Occurs if a factor (Z) that influences Y is correlated with X and left out of the regression.

Direction of Bias: Bias = beta_Z * Corr(X, Z). Positive Bias means OLS overestimates; Negative Bias means it underestimates.

2. Measurement Error

3. Simultaneous Causality (Reverse Causality)

Occurs if Y also affects X (e.g., price and quantity in a market).

4. Sample Selection Bias

Occurs if data is missing non-randomly (e.g., only surveyed people who survived a surgery).

9. Advanced Models

Binary Dependent Variables

When Y is 0 or 1 (e.g., yes/no, success/failure).

Instrumental Variables (IV)

Used to solve Endogeneity (when E[u|X] is not 0). Requires an Instrument (Z):

  1. Relevance: Z must be correlated with X (Corr(Z, X) != 0).
  2. Exogeneity: Z must affect Y only through its effect on X (Corr(Z, u) = 0).
beta_1_IV = Cov(Z, Y) / Cov(Z, X)

Panel Data (Fixed Effects)

Data tracking the same entities over multiple time periods.

Fixed Effects (FE): Controls for all unobserved factors that are constant for each entity (e.g., cultural traits of a country).

10. Time Series Analysis

Autocorrelation

When a variable is correlated with its own past values. Measured by Lagged variables (Y subscript t minus 1).

Stationarity

A series is stationary if its mean and variance do not change over time. Non-stationary series (like stock prices) often have Unit Roots or Stochastic Trends.

AR(p) and ADL(p, q) Models

Cointegration

If two non-stationary series move together in the long run, they are "Cointegrated."