Topic_3_Multivariate_OLS_Flashcards

(20 cards)

1
Q

What is the multivariate linear regression model?

A

Yi = β0 + β1X1i + β2X2i + … + βkXki + ui. It studies how multiple explanatory variables jointly affect one dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does the zero conditional mean assumption imply in the multivariate case?

A

E[ui | X1i, X2i, …, Xki] = 0. It ensures the error term is uncorrelated with all regressors, making OLS unbiased and consistent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the four key Least Squares assumptions in the multivariate model?

A

1) Zero conditional mean: E[ui | X1i, …, Xki] = 0; 2) Data are i.i.d.; 3) Large outliers are unlikely (finite fourth moments); 4) No perfect multicollinearity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the dummy variable trap?

A

Including all categories’ dummies (plus an intercept), which creates perfect multicollinearity and prevents OLS estimation. Solution: omit one base category.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is imperfect multicollinearity?

A

Regressors are highly, but not perfectly, correlated, making it difficult to distinguish individual effects.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Why is multicollinearity a problem?

A

It inflates the variances of estimated coefficients, yielding less precise estimates and smaller t-statistics.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is omitted variable bias (OVB)?

A

Bias that occurs when a relevant variable correlated with an included regressor is omitted and absorbed in the error term.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Which Least Squares assumption is violated when there is omitted variable bias?

A

The zero conditional mean assumption: E[ui | Xi] ≠ 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Should we include all variables to avoid omitted variable bias? Why (not)?

A

No. Include relevant, predetermined regressors. Irrelevant variables increase variance, and ‘bad controls’ bias causal interpretation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are ‘bad controls’?

A

Variables that are outcomes of the treatment or regressors of interest (e.g., occupation when estimating the effect of education on earnings).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a joint hypothesis?

A

A hypothesis imposing multiple coefficient restrictions simultaneously (e.g., H0: β3 = β4 = 0).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Which test is used to test a joint hypothesis?

A

The F-test, which compares the fit of restricted vs. unrestricted models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

When should the F-test be used instead of the t-test?

A

Use an F-test for two or more simultaneous restrictions; use a t-test for a single restriction on one coefficient.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Why can’t t-tests be used for joint hypotheses?

A

Because coefficients can be correlated; the F-test accounts for their covariance and tests joint significance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the conditional mean independence assumption?

A

Once controls are included, the expected error does not depend on regressors: E[ui | X1i, …, Xki] = 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the difference between a control variable and a variable of interest?

A

A variable of interest’s coefficient measures the causal effect of interest; controls isolate this effect by holding other influences constant.

17
Q

Do coefficients on control variables measure causal effects?

A

Not necessarily. They often capture associations used to adjust for confounding rather than direct causal effects.

18
Q

What are the main measures of fit in a multivariate regression model?

A

R², Adjusted R², and the Standard Error of the Regression (SER).

19
Q

When interpreting coefficients in log-linear models, what does β1 mean?

A

β1 × 100 gives the approximate percentage change in Y for a one-unit increase in X (if β1 is small).

20
Q

What does model specification in practice involve?

A

Choosing a base model grounded in theory, testing alternative specifications, and checking robustness of coefficients to ensure results are not driven by model choice.