Explore tens of thousands of sets crafted by our community.
Introduction to Item Response Theory
20
Flashcards
0/20
Item Response Theory (IRT)
IRT is a framework for designing, analyzing, and scoring tests, questionnaires, and similar instruments that measure abilities, aptitudes, or traits. It emphasizes the relationship between the latent trait under examination and the probability of a given response.
Latent Trait
In IRT, a latent trait is an unobservable characteristic or attribute such as ability or personality, which the test is designed to measure.
Item Characteristic Curve (ICC)
The ICC is a graph that plots the probability of a correct response to an item against the latent trait level of the examinee.
Item Difficulty (b-parameter)
The item difficulty parameter in IRT indicates the point on the latent trait scale at which an item has a 50% chance of being answered correctly. A higher value means that the item is more difficult.
Item Discrimination (a-parameter)
This is a measure of how well an item can differentiate between respondents with differing levels of the latent trait. A higher a-parameter indicates better discrimination.
Guessing Parameter (c-parameter)
In IRT, the guessing parameter accounts for the likelihood that a respondent will answer an item correctly by guessing. This is particularly relevant in multiple-choice items.
Theta (θ)
Theta is a symbol often used to represent the latent trait level of an individual in IRT models.
Three-Parameter Logistic Model (3PL)
The 3PL model in IRT incorporates three parameters: item difficulty (b), item discrimination (a), and guessing (c), to explain the probability of a correct response.
Two-Parameter Logistic Model (2PL)
The 2PL model is a version of IRT that includes only two parameters: item difficulty (b) and item discrimination (a), without considering the guessing factor.
One-Parameter Logistic Model (1PL)
Also known as the Rasch model, the 1PL assumes that all items have the same discrimination parameter (a) and only the difficulty (b) parameter varies among items.
Test Information Function (TIF)
TIF in IRT shows the amount of information that a test or a set of items provides across different levels of the latent trait. Higher information implies more precision in measuring the trait.
Assumption of Unidimensionality
This assumption in IRT states that a test measures a single latent trait or ability. All items on the test contribute to assessing this one dimension.
Assumption of Local Independence
In IRT, the assumption of local independence means that responses to items are independent of each other given the latent trait level of the individual.
Scaling in IRT
Scaling involves converting raw test scores onto a scale that represents the latent trait (usually theta) using an IRT model.
Differential Item Functioning (DIF)
DIF occurs when items on a test have different probabilities of a correct response for individuals of the same ability level but different subgroups (e.g., gender, ethnicity).
IRT Model Fit
Model fit in IRT examines the extent to which the data collected fits the chosen IRT model, indicating how well the model describes the performance of items.
Maximum Likelihood Estimation (MLE)
MLE in IRT is a statistical method used to estimate the parameters of the IRT model that maximize the likelihood of the observed set of responses.
Bayesian Estimation
Unlike MLE, Bayesian estimation in IRT involves prior distributions for parameters and uses Bayes' theorem to update parameter estimates given the data.
Expected a Posteriori (EAP) Estimation
EAP is a Bayesian estimation technique in IRT that produces point estimates of ability for examinees by taking a weighted average, where the weights are the posterior probabilities.
Standard Error of Measurement (SEM)
SEM in IRT is an index of the amount of error associated with an individual's estimated ability level, reflecting the precision of the estimate.
© Hypatia.Tech. 2024 All rights reserved.