Beruflich Dokumente
Kultur Dokumente
Validity &
Reliability
Presented By:
Sunil Kumar Saroha
Reliability of Test
Thereliability of a testrefers to stability of
measurement over time. When a person's data
entry skills are measured on two occasions
(with no special training in between), the two
sets of scores should be similar. Reliability is
often measured with areliability coefficient,
which is simply a correlation between sets of
scores from people who have been given the
test on two occasions (X = first time score on
the test, Y = second time score on the test)
Test-Retest
Test-retest - the same test is
given to the same people on
two occasions. The scores are
correlated, and if the reliability
coefficient is positive and
high, the test is reliable.
Example of TestRetest
Split-Half
Split-half- after being taken by a
sample, the answers to the test are
divided into two halves (e.g., the
odd-numbered versus the evennumbered items). Scores on each
half are correlated. If the test is
reliable, the scores on the two
Example of Split-Halt
Split-half- Each person has two scores, each on 1/2 of the
exam.
Score on even-numbered Score on odd-numbered
Student ID
items (30 possible)
items (30 possible)
1
23
24
2
30
29
3
26
25
4
19
20
5
24
23
Alternate Forms
Alternate forms- two versions of the
test are constructed, and given to
the same people on two occasions.
Scores on the two forms should show
a high positive reliabiilty coefficient
(correlation).
Student ID
1
2
3
4
5
Testing Validity
Construct validity
Content validity
Content validity - sometimes known as "face
validity." On the face of it, does the test or
inventory look like it measures what it is
supposed to measure? Does the content make
sense? An item asking about how gears work
would appear to measure mechanical ability.
Asking someone to indicate on a checklist
whether they are feeling happy or sad sounds
like a good way to find about that person's
Criterion validity
Concurrent Validity
Predictive Validity