Item response theory (IRT) represents an important innovation in the field of psychometrics. While now 50 years old – assuming the “birth” is the classic Lord and Novick (1969) text – it is still underutilized and remains a mystery to many practitioners. So what is item response theory, and why was it invented?
The Driver: Problems with Classical Test Theory
Classical test theory (CTT) is approximately 100 years old, and still remains commonly used because it is appropriate for certain situations, and it is simple enough that it can be used by many people without formal training in psychometrics. Most statistics are limited to means, proportions, and correlations. However, its simplicity means that it lacks the sophistication to deal with a number of very important measurement problems. Here are just a few.
- Sample dependency: Classical statistics are all sample dependent, and unusable on a different sample; results from IRT are sample-independent within a linear transformation (that is, two samples of different ability levels can be easily converted onto the same scale)
- Test dependency: Classical statistics are tied to a specific test form, and do not deal well with sparse matrices introduced by multiple forms, linear on the fly testing, or adaptive testing
- Weak linking/equating: CTT has a number of methods for linking multiple forms, but they are weak compared to IRT
- Measuring the range of students: Classical tests are built for the average student, and do not measure high or low students very well; conversely, statistics for very difficult or easy items are suspect
- Lack of accounting for guessing: CTT does not account for guessing on multiple choice exams
- Scoring: Scoring in classical test theory does not take into account item difficulty.
- Adaptive testing: CTT does not support adaptive testing in most cases.
So what is item response theory?
It is a family of mathematical models that try to describe how examinees respond to items (hence the name). These models can be used to evaluate item performance, because the description are quite useful in and of themselves. However, item response theory ended up doing so much more – namely, addressing the problems above.
The Foundation of Item Response Theory
The foundation of IRT is a mathematical model defined by item parameters. For dichotomous items (those scored correct/incorrect), each item has three parameters:
a: the discrimination parameter, an index of how well the item differentiates low from top examinees; typically ranges from 0 to 2, where higher is better, though not many items are above 1.0.
b: the difficulty parameter, an index of what level of examinees for which the item is appropriate; typically ranges from -3 to +3, with 0 being an average examinee level.
c: the pseudoguessing parameter, which is a lower asymptote; typically is focused on 1/k where k is the number of options.
These parameters are used to graphically display an item response function (IRF). An example IRF is on the right. Here, the a parameter is approximately, 1.0, indicating a fairly discriminating item. The b parameter is approximately -0.6 (the point on the x-axis where the midpoint of the curve is), indicating an easy item; examinees well below average would have a 60% chance of answering correctly. The c parameter is approximately 0.20, though the lower asymptote is obviously off the left of the screen.
What does this mean conceptually? We are trying to model the interaction of an examinee with the item, hence the name item response theory. Consider the x-axis to be z-scores on a standard normal scale. Examinees with higher ability are much more likely to respond correctly. Someone at +2.0 (97th percentile) has about a 94% chance of getting the item correct. Meanwhile, someone at -2.0 has only a 37% chance.
Building with the Basic Building Block
The IRF is used for several purposes. Here are a few.
- Interpreting and improving item performance
- Scoring examinees with maximum likelihood or Bayesian methods
- Form assembly, including linear on the fly testing (LOFT) and pre-equating
- Calculating the accuracy of examinee scores
- Development of computerized adaptive tests (CAT)
- Data forensics to find cheaters or other issues.
In addition to being used to evaluate each item individually, IRFs are combined in various ways to evaluate the overall test or form. The two most important approaches are the conditional standard error of measurement (CSEM) and the test information function (TIF). The test information function is higher where the test is providing more measurement information about examinees; if relatively low in a certain range of examinee ability, those examinees are not being measured accurately. The CSEM is the inverse of the TIF, and has the interpretable advantage of being usable for confidence intervals; a person’s score plus or minus 1.96 times the SEM is a 95% confidence interval for their score. The graph on the right shows part of the form assembly process in our FastTest platform.
One Big Happy Family
IRT is actually a family of models, making flexible use of the parameters. In some cases, only two (a,b) or one parameters (b) are used, depending on the type of assessment and fit of the data. If there are multipoint items, such as Likert rating scales or partial credit items, the models are extended to include additional parameters. Learn more about the partial credit situation here.
Where can I learn more?
For more information, we recommend the textbook Item Response Theory for Psychologists by Embretson & Riese (2000) for those interested in a less mathematical treatment, or de Ayala (2009) for a more mathematical treatment. If you really want to dive in, you can try the 3-volume Handbook of Item Response Theory edited by van der Linden, which contains a chapter discussing ASC’s IRT analysis software, Xcalibre.
Nathan Thompson, PhD
Latest posts by Nathan Thompson, PhD (see all)
- What is the Positive Manifold? - October 15, 2021
- Certification and Licensure – What is the difference? - October 10, 2021
- Certification Exams with Remote Proctoring: Best Practices - October 7, 2021