Documente Academic
Documente Profesional
Documente Cultură
VALIDITY
Does the test measure what it
is supposed to measure?
ANY GOOD TEST
SHOULD BE…
RELIABILITY
How consistent are the
✓ VALID results of the test?
✓ RELIABLE
✓ PRACTICAL PRACTICALITY
How practical is the test to
administer?
1. Validity
Know what is the purpose of the test & how the results will be used!
Ensuring the accuracy of the test & its desired interpretation
5 questions
Aina gets an Form 4 syllabus
‘A’ for a
subject in SPM
+ 35 questions
Form 5 syllabus
Domain Representativeness
What are the Are the number of items
domains/areas/parameters/ selected adequate/enough to
content being tested? measure students’
understanding?
B. Criterion validity
• The extent to which test results can be used to:
Predictive
• Infer/predict students' future performance
• Two measures obtained at different occasions (there is time interval)
Concurrent
• Estimate students' current performance
• Two measures obtained concurrently (no time interval between tests)
C. Construct validity
• The extent to which test results can be inferred to reflect a particular/various
constructs
• The legitimacy of interpretations made from students' test results to a construct
• Deals with human characteristics/quality
• Mathematical reasoning abilities, reading comprehension, creativity etc.
• Sociability, anxiety, leadership etc.
• Can be used to explain behaviour
E. Consequential
• Consequences of a test = social, psychological
• Effects of the test (intended & unintended) on students & teachers
• Positive (increased motivation) vs. negative (demotivated)
2. Reliability
A test is reliable when it produces consistent results
Does the test measure what it intends to measure in a reliable manner?
1st time
Score: 75
Measurement error
MEdSI When M.E. is less = The test is more reliable
2nd time
Score: 77
B. Equivalent-forms/Parallel-forms reliability
• Administering two forms of the test that measures the same contents,
domains or behaviours
• Are scores consistent over different forms?
• Forms of test are different but equivalent (e.g. Set 1, Set 2 etc.)
• Time interval should not be too long if achievement tests
C. Inter-rater reliability
• Consistency of judgement across different assessors
• Assessors use same tasks/procedures/standards
• Improved with clear scoring rules & training
• Correlation of scores between the different assessors
•
D. Internal consistency reliability
• The degree to which items in a test is consistent with one another
Split-half
• Administer test once
• Score odd & even items separately – as though dividing a test in two equal
halves
• Correlation of sub scores – estimates the reliability of a full assessment
Excercise
What do you want to measure? What will you use the results for?
Current level of students’ knowledge
in Form 4 Modern Mathematics
+ To monitor students’ ongoing progress
To improve instruction