Refreshers Flashcards

(49 cards)

1
Q

What is the purpose of studying the relationship between two quantitative variables?

A

To determine whether the variables are related, the strength and type of the relationship, and whether predictions can be made.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

A __________ is used to visually display the relationship between two quantitative variables.

A

Scatterplot

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

True / False: A scatterplot can show direction, form, strength, and outliers in a relationship.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does the correlation coefficient (r) measure?

A

The strength and direction of a linear relationship between two quantitative variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the range of the correlation coefficient r?
A) 0 to 1
B) –∞ to +∞
C) –1 to +1
D) –100 to +100

A

C

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

True / False: A correlation of r = 0 implies no relationship of any kind between two variables

A

False (it means no linear relationship)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

A positive correlation means that as one variable increases, the other variable __________.

A

Also increases

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the null and alternative hypotheses for a correlation test?

A

H₀: ρ = 0
H₁: ρ ≠ 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

When is the null hypothesis rejected in a correlation test?

A

When |r| is greater than the critical value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

True / False: Correlation implies causation

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Why can combining different subgroups in a dataset produce misleading correlations?

A

Because relationships within groups can be masked or reversed when data are combined

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is linear regression used for?

A

To model the relationship between variables, describe trends, and make predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The regression line is also called the line of __________

A

Best fit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

True / False: Regression should be performed even if the correlation is not significant

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is an independent variable (IV)?

A

The variable that causes or predicts changes in another variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a dependent variable (DV)?

A

The variable that is measured as the outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

In a study on sleep and test scores, which is the IV and which is the DV?

A

IV: Hours of sleep
DV: Test score

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

One way to identify the DV is to ask which variable occurs __________ in time.

A

Later

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is the strongest evidence for establishing causation?

A

Experimental evidence with replication.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What is the main purpose of linear regression?

A

To model a linear relationship and make predictions about one quantitative variable from another

21
Q

What method is used to find the regression line?

A

Least squares

22
Q

What is the “line of best fit”?

A

The line that best represents the overall trend of the data.

23
Q

Why are regression assumptions important?

A

Violated assumptions can invalidate the model and its conclusions

24
Q

What is a residual?

A

The vertical distance between an observed value and its predicted value

25
Name the five key assumptions of linear regression
1. Linearity 2. Normality of residuals 3. Equal variance of residuals (homoscedasticity) 4. Independence of residuals 5. No multicollinearity
26
What type of variable must the dependent variable (DV) be in linear regression?
Continuous
27
True / False: Multicollinearity affects the dependent variable directly
False
28
True / False: Linearity refers to the distribution of the residuals.
False
29
What are residuals in regression?
The differences between observed values and predicted values.
30
Unequal variance of residuals is called __________
Heteroscedasticity
31
What does the normality of residuals assumption require?
Residuals should be approximately normally distributed
32
What is an observed value in regression?
The actual value of the dependent variable collected from data (yᵢ)
33
Can residuals be negative?
Yes, if the predicted value is larger than the observed value (yᵢ – ŷᵢ < 0)
34
What is a predicted value in regression?
The value estimated by the regression model for a given independent variable (ŷᵢ)
35
What is the general form of a multiple linear regression equation?
Where Y = DV, Xs = IVs, βs = coefficients, e = residual 𝑌𝑖 = 𝛽0 + 𝛽1(𝑋1𝑖) + 𝛽2 (𝑋2𝑖) + 𝑒𝑖
36
What is leverage?
Extreme values in independent variable(s) (X) that can disproportionately affect regression results
37
What VIF value indicates no collinearity?
VIF = 1
38
What VIF value often indicates problematic collinearity?
VIF > 10 (though problems can occur with lower values)
39
What is repetitiveness in MLR?
Multiple measures of the same construct included as IVs.
40
T/F: Perfectly correlated IVs (r = 1) cause standard errors to be computable
False — regression fails; matrix cannot be inverted
41
Name three ways to detect collinearity.
1. Large changes in regression coefficients when adding/removing IVs 2. Unexpected signs of coefficients or large standard errors 3. Variance Inflation Factor (VIF)
42
Name two effects of high collinearity on regression results
Misleading significance of IVs (some may appear non-significant) Inflated standard errors of coefficients
43
What is the Variance Inflation Factor (VIF)?
A measure of how much the variance of a regression coefficient is inflated due to collinearity among IVs
44
How are outliers detected?
By examining studentized residuals
45
What is effect size?
A standardized measure of the magnitude of a difference or relationship, independent of sample size
46
How is Type I and Type II error expressed probabilistically?
Type I: 𝑃(Reject H₀ | H₀ true) = 𝛼 Type II: 𝑃(Fail to reject H₀ | H₀ false) = 𝛽
47
How do you calculate Cohen’s d for independent samples?
Cohen's d = (M_2 - M_1) ⁄ SDpooled Where, SDpooled = √((SD_1^2 + SD_2^2) ⁄ 2)
48
Can a “small” effect size still be important?
Yes — even small effects can have practical or cumulative significance, especially in large populations.
48
What is an a priori power analysis?
Estimating the minimum sample size needed to detect a true effect or confirm no significant difference, considering sample size, alpha, and effect size