Sie sind auf Seite 1von 23

RELIABILITY

• Reliability
- refers to the consistency of scores
obtained by the same person when re-
examined with the same test on
different occasions, or with different
sets of equivalent items, or under other
variable examining condition.
KINDS OF RELIABILITY
1. Test-Retest
2. Parallel Forms
3. Inter-rater
4. Split-Half
TEST - RETEST

• Same test
• Same examinee
• Differ in time. There’s a gap
ALTERNATE FORM RELIABILITY

• Same examinee
• Same time
• But differ in forms. There’s an alternate form/test.
• The two forms use different items; however, the rules used
to select items of a particular difficulty level are the same
INTER-RATER RELIABILITY

• Used to assess the degree to which


different judges or raters agree in their
assessment decision.
SPLIT HALF RELIABILITY

• It is obtained by splitting the items on a


questionnaire or test in half, computing a separate
score for each half, and then calculating the degree
of consistency between the two scores for a group of
participants.
SUMMARY

TEST
ADMINISTRATION TEST ITEMS
• Test-Retest Reliability • Split Half
• Alternate Form
Reliability

SCORING SYSTEM
• Inter-rater Reliablity
HOW TO IMPROVE RELIABILITY?

1. Quality of items; concise statements,


homogenous words (some sort of uniformity)
2. Adequate sampling of content domain;
comprehensiveness of items
3. Longer assessment – less distorted by chance
factors
4. Developing a scoring plan (esp. for subjective
items – rubrics)
5. Ensure Validity
VALIDITY
• Validity
- measures what it tends to measure
-its all about appropriateness
TYPES OF VALIDITY

1. Content Validity
2. Criterion-Related Validity
3. Construct Validity
CONTENT VALIDITY
• Involves an attempt to assess the content
of a test to assure it includes a
representative sample of all the questions
that could be asked
• Objectives and TOS
CRITERION – RELATED
VALIDITY
• How well a test corresponds with a particular
criterion provided by high correlations between a
test and a well-defined criterion measure

CONCURRENT PREDICTIVE
VALIDITY VALIDITY
CONCURRENT VALIDITY

• relationship between test scores and an


external criterion that is measured at
approximately the same time.

• Example: Reading readiness test might


be used to predict students’ achievement
in reading
PREDICTIVE VALIDITY

• Test scores may be obtained at one time


and the criterion measure may be obtained
in the future after an intervening event.

Example: Aptitude Test


CONSTRUCT VALIDITY
• Variable being measured by the test
(ano ang sinusukat ng test?)
• A construct is an unobservable trait that is known to exist.
Examples of constructs:
• IQ
• Anxiety
• Leadership Ability
• Motivation
• Self Esteem CONVERGENT DIVERGENT
VALIDITY VALIDITY
• Depression Same results Different results
FACTORS THAT CAN LOWER VALIDITY

- Unclear directions
- Difficult reading vocabulary and sentence structure
- Ambiguity in statements
- Inadequate time limits
- Inappropriate level of difficulty
- Poorly constructed test items
- Test items inappropriate for the outcomes being measured
- Tests that are too short
- Improper arrangement of items (complex to easy?)
- Identifiable patterns of answers
- Teaching
- Administration and scoring
- Nature of criterion
REMEMBER…

• A valid test is always reliable


but a reliable test is not necessarily valid
SUMMARY
THANK YOU!

Das könnte Ihnen auch gefallen