We show how nonlinear semi-supervised embedding algorithms popular for use with "shallow" learning techniques such as kernel methods can be easily applied to deep multi-layer architectures, either as a regularizer at the output layer, or on each layer of the architecture. Compared to standard supervised backpropagation this can give significant gains. This trick provides a simple alternative to existing approaches to semi-supervised deep learning whilst yielding competitive error rates compared to those methods, and existing shallow semi-supervised techniques. © Springer-Verlag Berlin Heidelberg 2012.
CITATION STYLE
Weston, J., Ratle, F., Mobahi, H., & Collobert, R. (2012). Deep learning via semi-supervised embedding. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 7700 LECTURE NO, 639–655. https://doi.org/10.1007/978-3-642-35289-8_34
Mendeley helps you to discover research relevant for your work.