Sum of Squared Errors (Definition)
Difference between Yi and Ŷ (observation and estimate)
Sum of Squared Regression (Definition)
Difference between Ŷ and Mean of Y (regression and best descriptive estimator)
SSR (#Degrees of Freedom)
k (# parameters of X estimated in the regression)
SSE (#Degrees of Freedom)
n-k-1 (N - estimators of X - intercept)
SST (#Degrees of Freedom)
(n-1)
Mean Squared of Regression (Formula)
MSR = SSR/k
Mean Squared of Error (Formula)
MSE = SSE/(n-k-1)
Squared Error of Estimate (SEE Formula)
SSE = √MSE
The lower, the more accurate the model is
F Test
F = MSR / MSE (testar a diferença entre a regressão em comparação com o erro)
DF @ K numerator (horizontal)
DF @ N-K-1 denominator (vertical)
Regression Assumptions
B0 (Intercept Test)
T-Test = (B1 est - B1 hipótese) / SB1
One-tail or Two-tails @ df = (n-k-1), as I am using error as a denominator
Sb1 = SEE / Sum of Sqaures of (Obs X - Mean X)
Dummy Variable
Y = b0 + b1*Dummy
Dummy = 0 or 1
If Dummy = 0, then Y = b0 = mean
If Dummy = 1, then Y = b0 + b1
Confidence Interval (Formula)
Interval = Ŷ ± T-Critical * Sf
Ŷ = Calculate using regression Sf = Std Error of Forecast = Sf = SEE² * [1 + 1/n * [(X-Mean)²/(n-1*Sx²)]
R² (Formula)
R² = SSR/SST = Measure of Fit
Regression Types
Multiple Regression Assumptions
F-statistic for Multiple (Hypothesis)
H0: B1 = B2 = B3 = 0
H1: At least one ≠ 0
One-Tailed Test @
DF Numerator = K = Horizontal
DF Denominator = (N-K-1) = Vertical
R² Adjusted (Formula)
Adj. R² = 1 - [(n-1)/(n-k-1)] * [1-R²]
Multicolinearity (Definition)
B1 e B2 t-tests are not relevant, but F-test is
Reason: Two IVs are highly correlated
Detection: ↑ R² and ↑ F-test, but ↓ B0
Correction: Omit one variable
Consequence: ↑ SE = ↓ F test
Heteroskedasticity (Definition)
Var of ε changes across observations
Unconditional: Var (ε) NOT correlated w/ IVs
Conditional: Var (ε) IS correlated w/ IVs
Correction:
Heteroskedasticity (Test)
Breusch Pagan Test (OH NO)
H0: NO conditional
H1: Conditional
Test = n * R²*ε @ Chi Squared Table
Regress the error on the IVs
Hansen Method (Definition)
Preferred if (i) SC or (ii) SC + Heteroskedasticity
Serial Correlation (Definition)
Test for Serial Correlation
Durbin Watson (Deutsche Welle)
H0: DW = 2 (No Correl)
H1: DW ≠ 2 (Correl)
Test = 2*(1-r) DF = K and N items
Correction: (i) Modified SEs,
(ii) Modify Regression Equation
(iii) Include seasonal term