<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
	<record>
		<datafield tag="980" ind1=" " ind2=" ">
			<subfield code="a">REPORT</subfield>
		</datafield>
		<datafield tag="970" ind1=" " ind2=" ">
			<subfield code="a">Alpa98/IDIAP</subfield>
		</datafield>
		<datafield tag="245" ind1=" " ind2=" ">
			<subfield code="a">Combined 5x2cv $F$-Test for Comparing Supervised Classification Learning Algorithms</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Alpaydin, Ethem</subfield>
		</datafield>
		<datafield tag="856" ind1="4" ind2="0">
			<subfield code="i">EXTERNAL</subfield>
			<subfield code="u">http://publications.idiap.ch/attachments/reports/1998/rr98-04.pdf</subfield>
			<subfield code="x">PUBLIC</subfield>
		</datafield>
		<datafield tag="088" ind1=" " ind2=" ">
			<subfield code="a">Idiap-RR-04-1998</subfield>
		</datafield>
		<datafield tag="260" ind1=" " ind2=" ">
			<subfield code="c">1998</subfield>
			<subfield code="b">IDIAP</subfield>
		</datafield>
		<datafield tag="500" ind1=" " ind2=" ">
			<subfield code="a">Submitted for publication</subfield>
		</datafield>
		<datafield tag="520" ind1=" " ind2=" ">
			<subfield code="a">Dietterich (1998) reviews five statistical tests proposing the 5x2cv t test for determining whether there is a significant difference between the error rates of two classifiers. In our experiments, we noticed that the 5x2cv t test result may vary depending on factors that should not affect the test and we propose a variant, the combined 5x2cv F test, that combines multiple statistics to get a more robust test. Simulation results show that this combined version of the test has lower Type I error and higher power than 5x2cv proper.</subfield>
		</datafield>
	</record>
</collection>