Effort in Low-Stakes Assessments: What Does It Take to Perform as Well as in a High-Stakes Setting? [0.03%]
低利害关系评价中的付出:在高利害关系设置中表现良好的要素是什么?
Yigal Attali
Yigal Attali
Performance of students in low-stakes testing situations has been a concern and focus of recent research. However, researchers who have examined the effect of stakes on performance have not been able to compare low-stakes performance to tru...
Evaluation of Measurement Instrument Criterion Validity in Finite Mixture Settings [0.03%]
有限混合情况下测量仪器评判有效性的评估
Tenko Raykov,George A Marcoulides,Tenglong Li
Tenko Raykov
A method for evaluating the validity of multicomponent measurement instruments in heterogeneous populations is discussed. The procedure can be used for point and interval estimation of criterion validity of linear composites in populations ...
Evaluating Rater Accuracy in Rater-Mediated Assessments Using an Unfolding Model [0.03%]
基于展开模型的评价者评分准确度评估研究
Jue Wang,George Engelhard Jr,Edward W Wolfe
Jue Wang
The number of performance assessments continues to increase around the world, and it is important to explore new methods for evaluating the quality of ratings obtained from raters. This study describes an unfolding model for examining rater...
Brandi A Weiss,William Dardick
Brandi A Weiss
This article introduces an entropy-based measure of data-model fit that can be used to assess the quality of logistic regression models. Entropy has previously been used in mixture-modeling to quantify how well individuals are classified in...
Reliability and Model Fit [0.03%]
可靠性与模型适配度
Leanne M Stanley,Michael C Edwards
Leanne M Stanley
The purpose of this article is to highlight the distinction between the reliability of test scores and the fit of psychometric measurement models, reminding readers why it is important to consider both when evaluating whether test scores ar...
An Approach to Scoring and Equating Tests With Binary Items: Piloting With Large-Scale Assessments [0.03%]
一种基于二元项目的评分和等值方法及其在大规模考试中的应用探索
Dimiter M Dimitrov
Dimiter M Dimitrov
This article describes an approach to test scoring, referred to as delta scoring (D-scoring), for tests with dichotomously scored items. The D-scoring uses information from item response theory (IRT) calibration to facilitate computations a...
Extracting Spurious Latent Classes in Growth Mixture Modeling With Nonnormal Errors [0.03%]
具有非正态误差的增长混合模型中抽取出伪潜在类别的方法
Kiero Guerra-Peña,Douglas Steinley
Kiero Guerra-Peña
Growth mixture modeling is generally used for two purposes: (1) to identify mixtures of normal subgroups and (2) to approximate oddly shaped distributions by a mixture of normal components. Often in applied research this methodology is appl...
Survey Satisficing Inflates Reliability and Validity Measures: An Experimental Comparison of College and Amazon Mechanical Turk Samples [0.03%]
草率作答会夸大测量的有效性和可靠性:以大学生和Amazon机械 Turks样本为例的实证研究
Tyler Hamby,Wyn Taylor
Tyler Hamby
This study examined the predictors and psychometric outcomes of survey satisficing, wherein respondents provide quick, "good enough" answers (satisficing) rather than carefully considered answers (optimizing). We administered surveys to uni...
Do Two or More Multicomponent Instruments Measure the Same Construct? Testing Construct Congruence Using Latent Variable Modeling [0.03%]
多组分指标是否测量相同的结构?利用潜变量模型检验结构一致性
Tenko Raykov,George A Marcoulides,Bing Tong
Tenko Raykov
A latent variable modeling procedure is discussed that can be used to test if two or more homogeneous multicomponent instruments with distinct components are measuring the same underlying construct. The method is widely applicable in scale ...
Tongyun Li,Hong Jiao,George B Macready
Tongyun Li
The present study investigates different approaches to adding covariates and the impact in fitting mixture item response theory models. Mixture item response theory models serve as an important methodology for tackling several psychometric ...