Hypothesis Testing for Generalized Thurstone Models

Anuran Makur, Japneet Singh
Proceedings of the 42nd International Conference on Machine Learning, PMLR 267:42730-42764, 2025.

Abstract

In this work, we develop a hypothesis testing framework to determine whether pairwise comparison data is generated by an underlying generalized Thurstone model $\mathcal{T}_F$ for a given choice function $F$. While prior work has predominantly focused on parameter estimation and uncertainty quantification for such models, we address the fundamental problem of minimax hypothesis testing for $\mathcal{T}_F$ models. We formulate this testing problem by introducing a notion of separation distance between general pairwise comparison models and the class of $\mathcal{T}_F$ models. We then derive upper and lower bounds on the critical threshold for testing that depend on the topology of the observation graph. For the special case of complete observation graphs, this threshold scales as $\Theta((nk)^{-1/2})$, where $n$ is the number of agents and $k$ is the number of comparisons per pair. Furthermore, we propose a hypothesis test based on our separation distance, construct confidence intervals, establish time-uniform bounds on the probabilities of type I and II errors using reverse martingale techniques, and derive minimax lower bounds using information-theoretic methods. Finally, we validate our results through experiments on synthetic and real-world datasets.

Cite this Paper


BibTeX
@InProceedings{pmlr-v267-makur25a, title = {Hypothesis Testing for Generalized Thurstone Models}, author = {Makur, Anuran and Singh, Japneet}, booktitle = {Proceedings of the 42nd International Conference on Machine Learning}, pages = {42730--42764}, year = {2025}, editor = {Singh, Aarti and Fazel, Maryam and Hsu, Daniel and Lacoste-Julien, Simon and Berkenkamp, Felix and Maharaj, Tegan and Wagstaff, Kiri and Zhu, Jerry}, volume = {267}, series = {Proceedings of Machine Learning Research}, month = {13--19 Jul}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v267/main/assets/makur25a/makur25a.pdf}, url = {https://proceedings.mlr.press/v267/makur25a.html}, abstract = {In this work, we develop a hypothesis testing framework to determine whether pairwise comparison data is generated by an underlying generalized Thurstone model $\mathcal{T}_F$ for a given choice function $F$. While prior work has predominantly focused on parameter estimation and uncertainty quantification for such models, we address the fundamental problem of minimax hypothesis testing for $\mathcal{T}_F$ models. We formulate this testing problem by introducing a notion of separation distance between general pairwise comparison models and the class of $\mathcal{T}_F$ models. We then derive upper and lower bounds on the critical threshold for testing that depend on the topology of the observation graph. For the special case of complete observation graphs, this threshold scales as $\Theta((nk)^{-1/2})$, where $n$ is the number of agents and $k$ is the number of comparisons per pair. Furthermore, we propose a hypothesis test based on our separation distance, construct confidence intervals, establish time-uniform bounds on the probabilities of type I and II errors using reverse martingale techniques, and derive minimax lower bounds using information-theoretic methods. Finally, we validate our results through experiments on synthetic and real-world datasets.} }
Endnote
%0 Conference Paper %T Hypothesis Testing for Generalized Thurstone Models %A Anuran Makur %A Japneet Singh %B Proceedings of the 42nd International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2025 %E Aarti Singh %E Maryam Fazel %E Daniel Hsu %E Simon Lacoste-Julien %E Felix Berkenkamp %E Tegan Maharaj %E Kiri Wagstaff %E Jerry Zhu %F pmlr-v267-makur25a %I PMLR %P 42730--42764 %U https://proceedings.mlr.press/v267/makur25a.html %V 267 %X In this work, we develop a hypothesis testing framework to determine whether pairwise comparison data is generated by an underlying generalized Thurstone model $\mathcal{T}_F$ for a given choice function $F$. While prior work has predominantly focused on parameter estimation and uncertainty quantification for such models, we address the fundamental problem of minimax hypothesis testing for $\mathcal{T}_F$ models. We formulate this testing problem by introducing a notion of separation distance between general pairwise comparison models and the class of $\mathcal{T}_F$ models. We then derive upper and lower bounds on the critical threshold for testing that depend on the topology of the observation graph. For the special case of complete observation graphs, this threshold scales as $\Theta((nk)^{-1/2})$, where $n$ is the number of agents and $k$ is the number of comparisons per pair. Furthermore, we propose a hypothesis test based on our separation distance, construct confidence intervals, establish time-uniform bounds on the probabilities of type I and II errors using reverse martingale techniques, and derive minimax lower bounds using information-theoretic methods. Finally, we validate our results through experiments on synthetic and real-world datasets.
APA
Makur, A. & Singh, J.. (2025). Hypothesis Testing for Generalized Thurstone Models. Proceedings of the 42nd International Conference on Machine Learning, in Proceedings of Machine Learning Research 267:42730-42764 Available from https://proceedings.mlr.press/v267/makur25a.html.

Related Material