Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

$β^{4}$-IRT: A New $β^{3}$-IRT with Enhanced Discrimination Estimation (2303.17731v1)

Published 30 Mar 2023 in cs.LG and stat.ML

Abstract: Item response theory aims to estimate respondent's latent skills from their responses in tests composed of items with different levels of difficulty. Several models of item response theory have been proposed for different types of tasks, such as binary or probabilistic responses, response time, multiple responses, among others. In this paper, we propose a new version of $\beta3$-IRT, called $\beta{4}$-IRT, which uses the gradient descent method to estimate the model parameters. In $\beta3$-IRT, abilities and difficulties are bounded, thus we employ link functions in order to turn $\beta{4}$-IRT into an unconstrained gradient descent process. The original $\beta3$-IRT had a symmetry problem, meaning that, if an item was initialised with a discrimination value with the wrong sign, e.g. negative when the actual discrimination should be positive, the fitting process could be unable to recover the correct discrimination and difficulty values for the item. In order to tackle this limitation, we modelled the discrimination parameter as the product of two new parameters, one corresponding to the sign and the second associated to the magnitude. We also proposed sensible priors for all parameters. We performed experiments to compare $\beta{4}$-IRT and $\beta3$-IRT regarding parameter recovery and our new version outperformed the original $\beta3$-IRT. Finally, we made $\beta{4}$-IRT publicly available as a Python package, along with the implementation of $\beta3$-IRT used in our experiments.

Summary

We haven't generated a summary for this paper yet.