10 Simple Linear Regression Flashcards

(22 cards)

1
Q

Predicted Values

A

Values of the dependent variable based upon estimated regression coefficients and the prediction about the value of the independent variable.

Y = b0 + b1*X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Confidence interval

A

Estimate a prediction interval around a predicted value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Standard error of the estimate (SEE)

A

An estimated regression does not describe the relationship between the dependent and independent variables perfectly.

The SEE is the standard deviation of the error term.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Confidence interval of Predicted Y

A

Y +- (t * s)

Where
s^2 = SEE^2 [ 1 + 1/n + (X-\bar(X))/(n-1)s^2

Degrees of freedom is n - 2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Functional forms

A

When the relationship between X and Y is not linear, fitting a linear model would result in biased predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Natural log transformation examples.

A

Log-lin: Log(Y) = bX
Lin-log: Y = bLog(X)
Log-Log = Log(Y) = bLog(X)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Simple linear regression

A

The variation in a dependent variable in terms of the variation in a single independent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Dependent variable

A

The variable whose variation is explained by the independent variable. Sometimes also referred to as the explained variable, endogenous variable, or the predicted variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Independent variable

A

The variable used to explain the variation of the dependent variable. Sometimes referred to as the explanatory variable, exogenous variable, or the predicting variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Ordinary Least of Squares formula

A

Y = b0 + b1X + e

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Regression Coefficients

A

b1 = Cov(X,Y)/\sigma^2
b0 = Y - \hat(b1)*X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Assumptions of Linear Regression

A

1) There is a linear relationship between the dependent and independent variables.
2) Variance of the error terms is constant (homoskedasticity)
3) Error terms are independently distributed.
4) Error terms are normally distributed.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Homoskedasticity

A

The case where prediction errors all have the same constant variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Heteroskedasticity

A

The variance of the error terms not being constant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Analysis of Variance (ANOVA)

A

SSE + SSR = SST
SST = \sum(Y -\bar(Y))^2
SSE = \sum(Y -\hat(Y))^2
SSR = \sum(\hat(Y) -\bar(Y))^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Mean Square Regression (MSR)

17
Q

Mean Square Error (MSE)

A

MSE = SSE (n-k-1)

18
Q

Coefficient of Determination (R^2)

A

R^2 measure the percentage of total variation in Y explained by the variation in X (SSR/SST)

19
Q

Standard Error of the Estimate (SEE)

A

Measures the accuracy of predicted values from the regression equation.

SEE = (SEE/(n -2))^.5 = MSE^.5

20
Q

The F-Statistic

A

Tests whether the independent variables explain the variation in the dependent variable.

Notes:
1) One-tailed test
2) 2 degrees of freedom

21
Q

F-Statistic Formula

A

F = MSR/MSE = (SSR/k)/(SSE/(n-k-1))

22
Q

Regression Coefficient t-Test

A

t = (\hat(b) - b)/s

where,
s= SEE/\sum(X-\bar(X))