<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
	<record>
		<datafield tag="980" ind1=" " ind2=" ">
			<subfield code="a">CONF</subfield>
		</datafield>
		<datafield tag="970" ind1=" " ind2=" ">
			<subfield code="a">Pinheiro_ICML_2014/IDIAP</subfield>
		</datafield>
		<datafield tag="245" ind1=" " ind2=" ">
			<subfield code="a">Recurrent Convolutional Neural Networks for Scene Labeling</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Pinheiro, Pedro H. O.</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Collobert, Ronan</subfield>
		</datafield>
		<datafield tag="856" ind1="4" ind2="0">
			<subfield code="i">EXTERNAL</subfield>
			<subfield code="u">http://publications.idiap.ch/attachments/papers/2014/Pinheiro_ICML_2014.pdf</subfield>
			<subfield code="x">PUBLIC</subfield>
		</datafield>
		<datafield tag="711" ind1="2" ind2=" ">
			<subfield code="a">31st International Conference on Machine Learning (ICML)</subfield>
			<subfield code="c">Beijing, China</subfield>
		</datafield>
		<datafield tag="773" ind1=" " ind2=" ">
			<subfield code="v">32</subfield>
			<subfield code="c">82-90</subfield>
		</datafield>
		<datafield tag="260" ind1=" " ind2=" ">
			<subfield code="c">2014</subfield>
			<subfield code="b">JMLR</subfield>
		</datafield>
		<datafield tag="856" ind1="4" ind2=" ">
			<subfield code="u">http://jmlr.org/proceedings/papers/v32/pinheiro14.html</subfield>
			<subfield code="z">URL</subfield>
		</datafield>
		<datafield tag="520" ind1=" " ind2=" ">
			<subfield code="a">The goal of the scene labeling task is to assign a class label to each pixel in an image. To ensure a good visual coherence and a high class accu- racy, it is essential for a model to capture long range (pixel) label dependencies in images. In a feed-forward architecture, this can be achieved simply by considering a sufficiently large input context patch, around each pixel to be labeled. We propose an approach that consists of a re- current convolutional neural network which al- lows us to consider a large input context while limiting the capacity of the model. Contrary to most standard approaches, our method does not rely on any segmentation technique nor any task- specific features. The system is trained in an end-to-end manner over raw pixels, and mod- els complex spatial dependencies with low infer- ence cost. As the context size increases with the built-in recurrence, the system identifies and cor- rects its own errors. Our approach yields state-of- the-art performance on both the Stanford Back- ground Dataset and the SIFT Flow Dataset, while remaining very fast at test time.</subfield>
		</datafield>
	</record>
</collection>