<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
	<record>
		<datafield tag="980" ind1=" " ind2=" ">
			<subfield code="a">REPORT</subfield>
		</datafield>
		<datafield tag="970" ind1=" " ind2=" ">
			<subfield code="a">Grandvalet_Idiap-RR-01-2009/IDIAP</subfield>
		</datafield>
		<datafield tag="245" ind1=" " ind2=" ">
			<subfield code="a">Support Vector Machines with a Reject Option</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Grandvalet, Yves</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Keshet, Joseph</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Rakotomamonjy, Alain</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Canu, Stéphane</subfield>
		</datafield>
		<datafield tag="856" ind1="4" ind2="0">
			<subfield code="i">EXTERNAL</subfield>
			<subfield code="u">http://publications.idiap.ch/attachments/reports/2008/Grandvalet_Idiap-RR-01-2009.pdf</subfield>
			<subfield code="x">PUBLIC</subfield>
		</datafield>
		<datafield tag="088" ind1=" " ind2=" ">
			<subfield code="a">Idiap-RR-01-2009</subfield>
		</datafield>
		<datafield tag="260" ind1=" " ind2=" ">
			<subfield code="c">2009</subfield>
			<subfield code="b">Idiap</subfield>
		</datafield>
		<datafield tag="771" ind1="2" ind2=" ">
			<subfield code="d">January 2009</subfield>
		</datafield>
		<datafield tag="520" ind1=" " ind2=" ">
			<subfield code="a">We consider the problem of binary classification where the classfier may abstain 
instead of classifying each observation. The Bayes decision rule for this setup, 
known as Chowâ€™s rule, is deï¬ned by two thresholds on posterior probabilities. 
From simple desiderata, namely the consistency and the sparsity of the classiï¬er, 
we derive the double hinge loss function that focuses on estimating conditional 
probabilities only in the vicinity of the threshold points of the optimal decision 
rule. We show that, for suitable kernel machines, our approach is universally 
consistent. We cast the problem of minimizing the double hinge loss as a quadratic 
program akin to the standard SVM optimization problem and propose an active set 
method to solve it efï¬ciently. We ï¬nally provide preliminary experimental results 
We consider the problem of binary classiï¬cation where the classiï¬er may abstain 
instead of classifying each observation. The Bayes decision rule for this setup, 
known as Chowâ€™s rule, is deï¬ned by two thresholds on posterior probabilities. 
From simple desiderata, namely the consistency and the sparsity of the classiï¬er, 
we derive the double hinge loss function that focuses on estimating conditional 
probabilities only in the vicinity of the threshold points of the optimal decision 
rule. We show that, for suitable kernel machines, our approach is universally 
consistent. We cast the problem of minimizing the double hinge loss as a quadratic 
program akin to the standard SVM optimization problem and propose an active set 
method to solve it efï¬ciently. We ï¬nally provide preliminary experimental results 
illustrating the interest of our constructive approach to devising loss functions.</subfield>
		</datafield>
	</record>
</collection>