what questions does validity answer
-is the test or measure meaningful?
-does it measure what it was supposed to measure?
reliable vs valid on dart board
-reliable: not on the bullseye but all darts are in the same place
-valid (+ technically reliable): all darts are on the bullseye, + also in the same place
types of research design validity
-internal validity
-external validity
internal validity
basically protocols that have to be followed
* high rigor
* high repeatability
* highly controlled
* low risk of bias
external validity
generalizability
* appliability
* transferability
* less controlled
measurement validity
accuracy
-does the instrument measure what it is supposed to?
face validity
subjective assessment of degree to which a test appears to measure what it is intended to
* most basic test of validity, not many papers use just this type anymore
* lacks operational definitions + comparison to reference standard
* does it pass the “eye test”?
* RED FLAG if something doesn’t pass this simple form of validity
red flag if something doesn’t pass which type of validity
face validity
-most basic test of validity
content validity
degree to which test items represent all relevant facets of the variable being measured
* theoretical concept, based on subjective opinion, so it can be difficult to measure directly (not as applicable)
* usually determined by a group of experts
construct validity
different from content validity because it is more SPECIFIC + emphasizes an operational definition
how do you assess construct validity?
-statistical analysis (simplest)- compare means
-analysis (simple)- correlation
-analysis (complex)- factor analysis
how to assess construct validity-
STATISTICAL ANALYSIS
-simplest
-compare means
-comparison of means via t-tests (p-value)
* should expect no differences between “like” groups (p is greater than 0.05)
* should expect significant differences between “unlike” groups (p is less than 0.05)
how to assess construct validity-
ANALYSIS (simple)
correlation
* convergent: strong association between “like” items or groups
* divergent (discriminant): weak or no association between “unlike” items or groups
* best when single measure exhibits both convergent + divergent
how to assess construct validity-
ANALYSIS (complex)
factor analysis
* used to identify which items on a questionnaire are highly correlated
* convergent items are grouped together (loaded), while divergent factors stay separate from one another (ex: function + symptoms factors contain different items)
* cross-loading is a problem where one item on a questionnaire loads onto multiple factors, so we don’t know what is actually being measured
cross-loading
a problem where one item on a questionnaire loads onto multiple factors, so we don’t know what is actually being measured
criterion-related validity
degree to which a measure of interest relates to a measure with established validity
2 types of criterion-related validity
-concurrent validity
-predictive validity
concurrent validity
comparing 2 measures that are taken at the same time
predictive validity
reflects the degree to which the results of a specific test can predict a future outcome
responsiveness
ability of a measure to detect change; clinical meaningfulness
example of responsiveness
-scores for a certain measure decrease for those that worsen
-scores for that same measure increase for those that improve
what 2 things are clinically practical
-criterion-related validity
-responsiveness
quantifying validity-
we normally report “r” for
association (relationship)
general guideline for validity with “r”
it is valid if r = 0.35 or greater