Boosted Backpropagation Learning for Training Deep Modular Networks - Robotics Institute Carnegie Mellon University

Boosted Backpropagation Learning for Training Deep Modular Networks

Alexander Grubb and J. Andrew (Drew) Bagnell
Conference Paper, Proceedings of (ICML) International Conference on Machine Learning, pp. 407 - 414, June, 2010

Abstract

Divide-and-conquer is key to building sophisticated learning machines: hard problems are solved by composing a network of modules that solve simpler problems (LeCun et al., 1998; Rohde, 2002; Bradley, 2009). Many such existing systems rely on learning algorithms which are based on simple parametric gradient descent where the parametrization must be predetermined, or more specialized per-application algorithms which are usually ad-hoc and complicated. We present a novel approach for training generic modular networks that uses two existing techniques: the error propagation strategy of backpropagation and more recent research on descent in spaces of functions (Mason et al., 1999; Scholkopf & Smola, 2001). Combining these two methods of optimization gives a simple algorithm for training heterogeneous networks of functional modules using simple gradient propagation mechanics and established learning algorithms. The resulting separation of concerns between learning individual modules and error propagation mechanics eases implementation, enables a larger class of modular learning strategies, and allows per-module control of complexity/regularization. We derive and demonstrate this functional backpropagation and contrast it with traditional gradient descent in parameter space, observing that in our example domain the method is signi cantly more robust to local optima.

BibTeX

@conference{Grubb-2010-10460,
author = {Alexander Grubb and J. Andrew (Drew) Bagnell},
title = {Boosted Backpropagation Learning for Training Deep Modular Networks},
booktitle = {Proceedings of (ICML) International Conference on Machine Learning},
year = {2010},
month = {June},
pages = {407 - 414},
keywords = {machine learning, boosting, optimization, backpropagation, deep learning, modular learning},
}