Home > Research > Publications & Outputs > Statistical Comparisons of Classifiers by Gener...

Links

View graph of relations

Statistical Comparisons of Classifiers by Generalized Stochastic Dominance

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published

Standard

Statistical Comparisons of Classifiers by Generalized Stochastic Dominance. / Jansen, Christoph; Nalenz, Malte; Schollmeyer, Georg et al.
In: Journal of Machine Learning Research, Vol. 24, 31.07.2023.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

Jansen, C, Nalenz, M, Schollmeyer, G & Augustin, T 2023, 'Statistical Comparisons of Classifiers by Generalized Stochastic Dominance', Journal of Machine Learning Research, vol. 24. <https://jmlr.org/papers/v24/22-0902.html>

APA

Jansen, C., Nalenz, M., Schollmeyer, G., & Augustin, T. (2023). Statistical Comparisons of Classifiers by Generalized Stochastic Dominance. Journal of Machine Learning Research, 24. https://jmlr.org/papers/v24/22-0902.html

Vancouver

Jansen C, Nalenz M, Schollmeyer G, Augustin T. Statistical Comparisons of Classifiers by Generalized Stochastic Dominance. Journal of Machine Learning Research. 2023 Jul 31;24.

Author

Jansen, Christoph ; Nalenz, Malte ; Schollmeyer, Georg et al. / Statistical Comparisons of Classifiers by Generalized Stochastic Dominance. In: Journal of Machine Learning Research. 2023 ; Vol. 24.

Bibtex

@article{51ffcff8c2b6455ba86fd6912942e3c0,
title = "Statistical Comparisons of Classifiers by Generalized Stochastic Dominance",
abstract = "Although being a crucial question for the development of machine learning algorithms, there is still no consensus on how to compare classifiers over multiple data sets with respect to several criteria. Every comparison framework is confronted with (at least) three fundamental challenges: the multiplicity of quality criteria, the multiplicity of data sets and the randomness of the selection of data sets. In this paper, we add a fresh view to the vivid debate by adopting recent developments in decision theory. Based on so-called preference systems, our framework ranks classifiers by a generalized concept of stochastic dominance, which powerfully circumvents the cumbersome, and often even self-contradictory, reliance on aggregates. Moreover, we show that generalized stochastic dominance can be operationalized by solving easy-to-handle linear programs and moreover statistically tested employing an adapted two-sample observation-randomization test. This yields indeed a powerful framework for the statistical comparison of classifiers over multiple data sets with respect to multiple quality criteria simultaneously. We illustrate and investigate our framework in a simulation study and with a set of standard benchmark data sets.",
author = "Christoph Jansen and Malte Nalenz and Georg Schollmeyer and Thomas Augustin",
year = "2023",
month = jul,
day = "31",
language = "English",
volume = "24",
journal = "Journal of Machine Learning Research",
issn = "1532-4435",
publisher = "Microtome Publishing",

}

RIS

TY - JOUR

T1 - Statistical Comparisons of Classifiers by Generalized Stochastic Dominance

AU - Jansen, Christoph

AU - Nalenz, Malte

AU - Schollmeyer, Georg

AU - Augustin, Thomas

PY - 2023/7/31

Y1 - 2023/7/31

N2 - Although being a crucial question for the development of machine learning algorithms, there is still no consensus on how to compare classifiers over multiple data sets with respect to several criteria. Every comparison framework is confronted with (at least) three fundamental challenges: the multiplicity of quality criteria, the multiplicity of data sets and the randomness of the selection of data sets. In this paper, we add a fresh view to the vivid debate by adopting recent developments in decision theory. Based on so-called preference systems, our framework ranks classifiers by a generalized concept of stochastic dominance, which powerfully circumvents the cumbersome, and often even self-contradictory, reliance on aggregates. Moreover, we show that generalized stochastic dominance can be operationalized by solving easy-to-handle linear programs and moreover statistically tested employing an adapted two-sample observation-randomization test. This yields indeed a powerful framework for the statistical comparison of classifiers over multiple data sets with respect to multiple quality criteria simultaneously. We illustrate and investigate our framework in a simulation study and with a set of standard benchmark data sets.

AB - Although being a crucial question for the development of machine learning algorithms, there is still no consensus on how to compare classifiers over multiple data sets with respect to several criteria. Every comparison framework is confronted with (at least) three fundamental challenges: the multiplicity of quality criteria, the multiplicity of data sets and the randomness of the selection of data sets. In this paper, we add a fresh view to the vivid debate by adopting recent developments in decision theory. Based on so-called preference systems, our framework ranks classifiers by a generalized concept of stochastic dominance, which powerfully circumvents the cumbersome, and often even self-contradictory, reliance on aggregates. Moreover, we show that generalized stochastic dominance can be operationalized by solving easy-to-handle linear programs and moreover statistically tested employing an adapted two-sample observation-randomization test. This yields indeed a powerful framework for the statistical comparison of classifiers over multiple data sets with respect to multiple quality criteria simultaneously. We illustrate and investigate our framework in a simulation study and with a set of standard benchmark data sets.

M3 - Journal article

VL - 24

JO - Journal of Machine Learning Research

JF - Journal of Machine Learning Research

SN - 1532-4435

ER -