<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
	<record>
		<datafield tag="980" ind1=" " ind2=" ">
			<subfield code="a">ARTICLE</subfield>
		</datafield>
		<datafield tag="970" ind1=" " ind2=" ">
			<subfield code="a">Do_JMLR_2012/IDIAP</subfield>
		</datafield>
		<datafield tag="245" ind1=" " ind2=" ">
			<subfield code="a">Regularized Bundle Methods for Convex and Non-Convex Risks</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Do, Trinh-Minh-Tri</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Artieres, Thierry</subfield>
		</datafield>
		<datafield tag="856" ind1="4" ind2="0">
			<subfield code="i">EXTERNAL</subfield>
			<subfield code="u">http://publications.idiap.ch/attachments/papers/2013/Do_JMLR_2012.pdf</subfield>
			<subfield code="x">PUBLIC</subfield>
		</datafield>
		<datafield tag="773" ind1=" " ind2=" ">
			<subfield code="p">Journal of Machine Learning Research</subfield>
			<subfield code="v">13</subfield>
			<subfield code="c">3539-3583</subfield>
		</datafield>
		<datafield tag="260" ind1=" " ind2=" ">
			<subfield code="c">2012</subfield>
		</datafield>
		<datafield tag="520" ind1=" " ind2=" ">
			<subfield code="a">Machine learning is most often cast as an optimization problem.
Ideally, one expects a convex objective function to rely on efficient
convex optimizers with nice guarantees such as no local optima. Yet,
non-convexity is very frequent in practice and it may sometimes be
inappropriate to look for convexity at any price. Alternatively one
can decide not to limit a priori the modeling expressivity to models
whose learning may be solved by convex optimization and rely on
non-convex optimization algorithms. The main motivation of this work
is to provide efficient and scalable algorithms for non-convex
optimization. We focus on regularized unconstrained optimization
problems which cover a large number of modern machine learning
problems such as logistic regression, conditional random fields, large
margin estimation, etc. We propose a novel algorithm for minimizing a
regularized objective that is able to handle convex and non-convex,
smooth and non-smooth risks. The algorithm is based on the cutting
plane technique and on the idea of exploiting the regularization term
in the objective function. It may be thought as a limited memory
extension of convex regularized bundle methods for dealing with convex
and non convex risks. In case the risk is convex the algorithm is
proved to converge to a stationary solution with accuracy ε with a
rate O(1/λε) where λ is the regularization parameter of the objective
function under the assumption of a Lipschitz empirical risk. In case
the risk is not convex getting such a proof is more difficult and
requires a stronger and more disputable assumption. Yet we provide
experimental results on artificial test problems, and on five standard
and difficult machine learning problems that are cast as convex and
non-convex optimization problems that show how our algorithm compares
well in practice with state of the art optimization algorithms.</subfield>
		</datafield>
	</record>
</collection>