Week 5: Validity Flashcards

(65 cards)

1
Q

Why is it not strictly accurate to talk about the validity of a test (hint: one test could be used in more than one context)?

A

It is not strictly accurate to talk about the validity of a test because test scores are able to be used and interpreted in more than one way, therefore making validity evaluations more about the context in which you are using a test, along with the persuasiveness of the test in the context measuring what it is said it is.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are constructs?

A

Constructs are unobservable, underlying, hypothetical traits or characteristics that we can try and measure indirectly using tests.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is construct underrepresentation?

A

The part of variation in the underlying construct being measured that is not captured.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is construct irrelevant variance?

A

The part of variation that a measure captures that is not related to the construct.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What’s the difference between content and face validity?

A

There is no true difference between content and face validity – the only difference is that content validity is often based on the opinions of experts whereas face validity tends to be based on the opinion of the person taking the test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Give your own example of a testing situation where content validity would be of central importance.

A

For a competence test – a hazard perception competence test for example

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How could I create a university examination that had great empirical validity but poor content validity?

A

By including questions that successfully predict students understanding of the course, without being about the content from the lectures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why would it not be okay for the PSYC3020 quizzes to have poor content validity (even if they could reliably tell apart high and low scoring students)?

A

Because it isn’t fair

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How could you go about evaluating content validity?

A

By asking a group of experts their opinions as to if the test measures what it is supposed to.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

List five types of validity.

A

Content validity, face validity, convergent validity, discriminant validity, and criterion validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does hypothesis testing have to do with evaluating the validity of a test?

A

To design an empirical validation study, hypotheses need to be created regarding how measures ought to perform if it is valid.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Is reliability necessary for validity?

A

Yes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Is validity necessary for reliability?

A

No

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Explain why reliability is necessary for validity but validity is not necessary for reliability.

A

It is problematic if there is no reliability, as there is nothing stable enough to be valid or not valid, however, reliability is not affected by validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Give two examples of things that might restrict the range of scores in a test and indicate what influence this could have on the validity coefficient.

A

Non-random attrition and self selection

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Give an example of how self-selection during participant recruitment might restrict the range of test scores.

A

Self-selection may restrict the range of test scores, as there is a possibility for there not to be a distribution which mimics the real world (missing a part of the population)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

How big does a validity coefficient have to be for a test to be considered valid?

A

The required magnitude depends entirely on the context – there is no specific cutoff.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is criterion validity?

A

Criterion validity is the evidence that a test score corresponds to an accurate measure of interest (the criterion).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Give your own example of criterion validity.

A

Giving a new test to 50 surgeons, get them to do it, find real world patient outcomes for each surgeon, and compare the test score to real life.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Give three examples of criterion measures that could be used to evaluate the criterion validity of relevant tests.

A

GPA at end of 1st year to measure university admissions test, supervisors ratings of job performance to measure clerical aptitude, panel rating of creativity displayed in artistic products to measure creative thinking.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is the criterion variable used to evaluate criterion validity?

A

The criterion is the standard against which the test is evaluated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What is the method of contrasted groups?

A

An approach to criterion validity, where it is determined if test scores of groups of people vary as expected.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What is criterion contamination?

A

Where the criterion used to assess the validity of a test is pre-determined by the test, thereby undermining the logic of criterion validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Give an example of criterion contamination when evaluating criterion validity.

A

Validating a test of schizophrenic people by seeing if it can tell those diagnosed from those not, and then discover that those people have been diagnosed using the same test looking to be validated (circular).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Does it matter if a criterion used to validate a test is not that reliable? Why?
Yes – because the reliability limits the size of how big the validity coefficient can be.
26
What’s the relationship between the reliability of the test/criterion and the validity coefficient?
If the reliability of the test is low then there is limited chance for the validity coefficient to be high or even accurate.
27
What is concurrent validity?
Concurrent validity is a subset of criterion validity where test and criterion are measured at the same time.
28
Give an example of concurrent validity.
If the test and criterion are administered in the same session
29
What is predictive validity?
Predictive validity is a subset of criterion validity where the test is trying to predict what the criterion will be at some future time
30
Give an example of predictive validity.
Waiting a period of time between obtaining the test score and criterion score.
31
What is incremental validity?
Incremental validity is how much each individual predictor adds to predicting the criterion in addition to the effect of other predictors – it is a type of criterion validity, and can be predictive or concurrent.
32
Give an example of incremental validity.
Measuring sensation-seeking, fear or heights, and susceptibility to peer pressure when predicting if someone would do a bungee jump.
33
What are the different ways you could go about testing the empirical validity of a test?
Criterion validity, convergent validity, construct validity, discriminant validity, measuring expected developmental changes, experimental effects and internal structure.
34
Give examples of convergent and discriminant validity.
An example of convergent validity is if test scores correlate with other measures of the same (or similar) thing. An example of discriminant validity is if test scores do not correlate very highly with measures that you would expect them not to correlate with.
35
How can we evaluate the validity of a test by inspecting its internal structure?
We can investigate the predictions about whether the internal structure of our test maps onto the internal structure of our construct using factor analysis.
36
What is factor analysis?
Factor analysis uses mathematical techniques to group items into clusters (factors or components) on the basis of how much they correlate with each other.
37
Explain what factor analysis does.
Factor analysis picks out the distinct clusters of items in our data based on their inter-correlations.
38
What is construct validity?
Whether or not the test measures the construct it is meant to.
39
True or False: Construct underrepresentation refers to the variance in a construct that does not covary with test scores.
True
40
True or False: Construct irrelevant variance refers to the variance in a test that does not covary with the construct of interest.
True
41
True or False: If the variances of a test and the construct it is attempting to measure overlap substantially, then the test is likely to have a high degree of validity.
True
42
True or False: If the variances of a test and the construct it is attempting to measure only overlap by a small degree, then the test is likely to have low reliability.
False - it is likely to have low validity
43
True or False: Fifty participants take a new test and then complete a questionnaire in which they rate the perceived validity of the test on a number of quantitatively-scored scales. We can use the rating data obtained from this questionnaire to evaluate the face validity of the test.
True
44
True or False: Face validity is ONLY really useful for persuading people to do the test in the first place.
False
45
True or False: Having good face validity is always an advantage for a test.
False
45
True or False: One useful thing about high face validity is that it might help to persuade people to do the test in the first place.
True
46
True or False: When students complain that, in a course examination, a lecturer did not ask any questions from a particular presentation, they are effectively complaining that the examination has poor content validity.
True
47
True or False: It is possible to have a test with excellent empirically-based support for its validity without having any content validity.
True
48
True or False: One way of evaluating a test’s content validity would be to ask a group of experts to complete a test and then examine their performance on each item for its validity.
False – their opinion hasn’t been asked which is key for content validity
49
True or False: One way of evaluating a test’s content validity would be to ask a group of experts to complete a test and then ask their opinions on whether the test content adequately covered the domain it was designed to cover.
True
50
True or False: If a test is reliable then we can infer that it must also be valid.
False
51
True or False: If we want to calculate a correlation involving a test and a validating measure, then any restrictions of range on the variables could be a problem.
False
52
True or False: If the reliability of both a test and a criterion measure is high, then this means the correlation between them should also be high.
False
53
True or False: Factors that may affect a predictive validity coefficient include certain types of people dropping out of the sample between the original test and when the criterion is measured.
True
54
True or False: You can evaluate the incremental validity of a test by seeing whether it can predict some relevant criterion measure IN ISOLATION from other measures.
False – it is in addition to other measures
55
True or False: Incremental validity refers to the total validity achieved by combining multiple tests to predict a single outcome.
False – referring specifically to what happens when you add an item to test battery, not the overall battery together
56
True or False: A zero correlation between a test score and a relevant validating scale for that test always means that the test is not valid.
False
57
True or False: Examining whether developmental changes occur as expected is one way of evaluating the validity of a test.
True
58
True or False: Examining whether developmental changes occur as expected is one way of evaluating the content validity of a test.
False
59
True or False: If we had an established intervention known to reduce state anxiety, then we potentially could use this to test the validity of a new measure of state anxiety.
True
60
True or False: Relating test scores to the framework of a theoretical construct is one method of evaluating the validity of a test.
True
61
True or False: A factor analysis involves mathematically grouping items according to their intercorrelations.
True
62
True or False: A factor analysis involves mathematically grouping items according to the similarity of the item content.
False
63
True or False: Face validity can be regarded as a type of construct validity.
False
64
True or False: Convergent validity evidence can be used to evaluate construct validity.
True