Give the Simple Linear Regression function…
Yb(x) = b0 + b1x + error
What is the formula for the Y-intercept of the regression line?
b0 = y median - (b1 * x median)
What is the formula for the slope of the regression line?
b1 = r * (sy / sx)
What is the goal of the simple regression function?
To achieve as small error as possible.
What is the formula for the error of the simple linear regression function?
Transpose the simple linear regression function with respect to E
Error = Yb(x) - b0 - b1x
What does MSE find?
The average error between the regression line and the data points.
Why do we square the errors for MSE?
What is the advantage of MSE?
What is the goal of MSE? What must we tune to obtain this?
What is the Least Square Method?
What are the 4 steps to the Least Squares Method?
What do we compare the Least Squares Method against?
The mean of Y, which is the most basic regression line.
What are the 2 types of variations encountered in LSM?
Unexplained Variations - The error that can’t be explained by the independent variable.
Explained Variations - The error that can be explained by the independent variable.
How do we calculate the Total Variation of LSM?
Total Variation = Unexplained variation + explained variation
What is the Sum of Squared Residual (SSR)?
Measures the level of variance in the error of a regression model.
What is the Sum of Squared Error (SSE)?
Measures the difference between the predicted values and the error mean (MSE).
What is the coefficient of determination?
Value between 0 and 1 that informs us of the prediction quality of the model.
1 is better.
What is the formula for R^2?
R^2 = 1 - (SSE/TSS)
In regression formulas, what is omega? When do we care about it?
What is the difference between Prediction and Interpretation?
Prediction: Using model to predict a value. We focus on performance metrics, not omega. Be careful about model becoming black box.
Interpretation: Using model to gain insight into data. We focus on parameters toe stablish feature impact.