How is Logistic regression similar to Multiple Regression?
* Like multiple regression, the prediction equation includes a linear combination of the predictor variables
What does Logistic Regression (AKA Logit Regression) enable researchers to achieve?
Logistic regression allows one to:
What is Binomial or Binary Logistic Regression?
Binomial (or Binary) logistic regression is a form of regression used when the single dependent variable is dichotomous, even though the independent variables may be of any type
What are some key terms to consider in Logistic Regression?
Although Binomial logistic regression is relatively free of restrictions, there are some limitations to be aware of. What are these?
What is a Logit Variable?
A Logit Variable is a natural log of the odds of the dependent occurring or not
When does Logistic regression apply Maximum Likelihood Estimation (MLE) ?
What is the difference between Maximum Likelihood Estimation (MLE) & Ordinary Least Square (OLS) estimation?
Because logistic regression calculates changes in the log odds of the dependent whereas OLS regression calculates changes in the dependent variable itself
Why does Howell (2002) favour Logistic Regression over alternatives to Logistic analysis such as Standard Multiple Regression (SMR) and Discriminant Function Analysis (DFA)?
With a dichotomous dependent variable SMR only provides a fairly good estimate if the percentage of improvement scores don’t fall below 20% or above 80% across all values, the rest of the time is not a wise choice. *DFA requires more stringent assumptions to be met & may produce probabilities out of the range being investigated, that is 0 to 1 (so not good)
What visual representation does Howell (2002) favour over a straight line?
What are the 2 steps required to calculate the probabilities?
NB: This aspect of analysis is sometimes known as a link function within statistics.
What kind of research questions can logistic regression address?
Assumption requirements of Logistic Regression vary according to the text you read, what do Hair, Black, Babin and Anderson (2011) suggest an advantage of logistic regression (LR)?
Hair, Black, Babin and Anderson (2011) suggest an advantage of logistic regression (LR) is, the lack of assumptions:
Assumption requirements of Logistic Regression vary according to the text you read, what would Pallant (2011) suggest?
Pallant (2011) would suggest you check sample size, multicollinearity and deal with any outliers by inspecting scatter plots if you have problems with goodness of fit in your model.
Assumption requirements of Logistic Regression vary according to the text you read, what would Andy Field (2013) recommend?
Field would suggest you check linearity of the relationship with the log of the outcome variable, check for large standard errors and over dispersion (caused by violating the assumption of independence).
Tell me a little more about how Pallant would ensure assumptions are met for Logistic Regression
It is necessary to test Goodness of Fit for Logistic Regression Models. How do we do this?
Hosmer and Lemeshow goodness of fit test computes a chi-square statistic using observed and expected frequencies. This test evaluates whether the model’s estimates, fit the data well. We want this test to be NOT significant, that is >.05.
What is alternative test of Goodness of Fit?
What is alternative test of Goodness of Fit?
An alternative test for goodness of fit of the model is the omnibus tests of model coefficients. However, it tests whether the model that has all predictors included, is significantly different to the model with just the intercept.
I have heard that sample size and interpretation of results are of particular importance in relation to goodness of fit. Why so?
In Logistic Regression, what do large R2 (R squared) values indicate?
In logistic regression larger R2 values indicate that more of the variation is explained by the model, to a maximum of one. However, for regression models with a categorical dependent variable, it is not possible to compute a single R2 statistic that has all of the characteristics of R2 in the linear regression model, so these alternative approximations are computed instead.
What does the Log Likelihood Statistic indicate in Logistic Regression?
Why is the classification table the most important part of estimating the probability of predicting the outcome in Logistic Regression?
*The classification table indicates the practical results of the model to the researcher, i.e. how much is predicted
How do we interpret logistic regression coefficients?
* If an acceptable model is found the Wald test statistic for each predictor is evaluated