TY - JOUR
T1 - Effects of response format on achievement and aptitude assessment results
T2 - multi-level random effects meta-analyses
AU - Breuer, Sonja
AU - Scherndl, Thomas
AU - Ortner, Tuulia M.
PY - 2023/5/3
Y1 - 2023/5/3
N2 - Psychological achievement and aptitude tests are fundamental elements of the everyday school, academic and professional lives of students, instructors, job applicants, researchers, and policymakers. In line with growing demands for fair psychological assessment tools, we aimed to identify psychometric features of tests, test situations, and test-taker characteristics that may contribute to the emergence of test bias. Multilevel random effects meta-analyses were conducted to estimate mean effect sizes for differences and relations between scores from achievement or aptitude measures with open-ended (OE) vs. closed-ended (CE) response formats. Results from 102 primary studies with 392 effect sizes revealed positive relations between CE and OE assessments (mean r = 0.67, 95% CI [0.57; 0.76]), with negative pooled effect sizes for the difference between the two response formats (mean dav = -0.65; 95% CI [-0.78; -0.53]). Significantly higher scores were obtained on CE exams. Stem-equivalency of items, low-stakes test situations, written short answer OE question types, studies conducted outside the US and before the year 2000, and test-takers’ achievement motivation and sex were at least partially associated with smaller differences and/or larger relations between scores from OE and CE formats. Limitations and the results’ implications for practitioners in achievement and aptitude testing are discussed.
AB - Psychological achievement and aptitude tests are fundamental elements of the everyday school, academic and professional lives of students, instructors, job applicants, researchers, and policymakers. In line with growing demands for fair psychological assessment tools, we aimed to identify psychometric features of tests, test situations, and test-taker characteristics that may contribute to the emergence of test bias. Multilevel random effects meta-analyses were conducted to estimate mean effect sizes for differences and relations between scores from achievement or aptitude measures with open-ended (OE) vs. closed-ended (CE) response formats. Results from 102 primary studies with 392 effect sizes revealed positive relations between CE and OE assessments (mean r = 0.67, 95% CI [0.57; 0.76]), with negative pooled effect sizes for the difference between the two response formats (mean dav = -0.65; 95% CI [-0.78; -0.53]). Significantly higher scores were obtained on CE exams. Stem-equivalency of items, low-stakes test situations, written short answer OE question types, studies conducted outside the US and before the year 2000, and test-takers’ achievement motivation and sex were at least partially associated with smaller differences and/or larger relations between scores from OE and CE formats. Limitations and the results’ implications for practitioners in achievement and aptitude testing are discussed.
KW - response format
KW - achievement and aptitude assessment
KW - open-ended
KW - closed-ended
KW - meta-analysis
UR - http://www.scopus.com/inward/record.url?scp=85160254052&partnerID=8YFLogxK
UR - https://pubmed.ncbi.nlm.nih.gov/37153364/
UR - https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10154931/
UR - https://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=univsalzburg_wosstarter&SrcAuth=WosAPI&KeyUT=WOS:000984742800003&DestLinkType=FullRecord&DestApp=WOS
UR - https://www.mendeley.com/catalogue/baf09198-6bfb-3600-ac33-e4a72a730068/
U2 - 10.1098/rsos.220456
DO - 10.1098/rsos.220456
M3 - Review article
C2 - 37153364
SN - 2054-5703
VL - 10
JO - Royal Society Open Science
JF - Royal Society Open Science
IS - 5
M1 - 220456
ER -