Three validities for causation
Define construct validity
Whether the construct is operationalised through the msmt
How much the concept is reflected in the msmt used
Variable true to itself
E.g. Wealth~Hrs worked = low; Wealth~ #assets and annual pay = high
Define reliability
The replicability of the msmt tool used
The accuracy and precision of the msmt tool used
E.g. measuring height using 1m stick (low) or 1m ruler (high)
Types of reliability
Define test-retest reliability
Correlation b/w scores at two time points
When a test taken multiple times produce the same results in individuals
Cons: Susceptible to learning
Cons: Works for static, not dynamic constructs
Define split-half reliability
Correlation b/w scores in both halves of experiment
When individuals tests’ are halved and the correlation between the two halves are measured.
Cons: Overtesting, test length
Define parity reliability
Correlation b/w scores in odd and even trials
When individuals tests’ are seperated into odd and even trials and the correlation between the two halves are measured.
Define criterion validity
How well the test reflects the criterion it aims to test
Empirical validity
E.g. How well the simulated surgery scores reflect doctors’ actual surgery outcomes. Testing for actual surgery outcomes.
Define content validity
How well the test appears to reflect the construct
Opinion validity
E.g. How much of the lecture content is present during the exam. May test knowledge of course through tutorial content only, but not wholly content-valid.
Define convergent validity
How well other correlates of the construct correlate with the test
Do the scores of related correlates converge?
E.g. How much Psyc3020 and stats grades correlate with Psyc3042 grades
Define divergent validity
How well other tests of the construct do NOT correlate with the test
Do the scores of non-related concepts diverge in magnitude?
E.g. How much dep scores correlate with anxiety - shouldnt correlate
How much hrs studying reflects alcohol consumption - should not correlate
Define common method variance
It is common for the methods used to show a variance in results when the construct should not.
The variance we see in test results that shouldn’t be present in the construct we try to measure
E.g. self-reported alcohol inebriation (“how drunk do you feel”) vs. # pints drunk vs. BAC vs. metabolism (BAC over time). Each test reflects different systematic variances of the method, not the results.
What are multi-trait, multi-method techniques?
Measuring convergent and divergent validity over multiple methods (self-report, observation, behavioural msmts, etc.)
Usually 2 constructs with 2 methods,
E.g. Assertiveness and aggression through self-report and observation
Define internal validity
Whether the msmt reflects a causal relationship between constructs
Variable true to its relationship
Failures of internal validity
Define spurious scoring
Coincidence
E.g. US Industrial manufacturing doctorates awards ~ US pc mozzarella cheese consumption, r=0.9
Define scoring and data analysis errors (internal validity)
The analysis misconstrues the results; misinterpretation.
E.g. A vaccine decreases fatalities in a disease from 23 in 1000 to 9 in 1000. Is this a 60% decrease or a decrease of 0.014%?
Define reverse causation (internal validity)
Assume X causes A when A causes X
E.g. Assume married people are happier because they married; actually happier people are the ones who marry.
Determining causation through experimentation
Define mediation
A -> C because A -> B -> C
Mediation vs mechanism of causation
Mediation is through another independent construct, whereas the mechanism doesn’t
E.g. mediation: guns cause aggression through feeling threatened
E.g. mechanism: studying improves grades because studying is the process of improving understanding and learning, and thus grades.
Measuring mediation
Full mediation
Partial mediation