MIT CogNet, The Brain Sciences ConnectionFrom the MIT Press, Link to Online Catalog
SPARC Communities
Subscriber : Stanford University Libraries » LOG IN

space

Powered By Google 
Advanced Search

 

Robust Neural Network Regression for Offline and Online Learning

 Thomas Briegel and Volker Tresp
  
 

Abstract:
We replace the commonly used Gaussian noise model in nonlinear regression by a more flexible noise model based on the Student -distribution. The degrees of freedom of the -distribution can be chosen such that as special cases either the Gaussian distribution or the Cauchy distribution are realized. The latter is commonly used in robust regression. Since the -distribution can be interpreted as being an infinite mixture of Gaussians, parameters and hyperparameters such as the degrees of freedom of the -distribution can be learned from the data based on an EM-learning algorithm. We show that modeling using the -distribution leads to improved predictors on real world data sets. In particular, if outliers are present, the -distribution is superior to the Gaussian noise model. In effect, by adapting the degrees of freedom, the system can "learn" to distinguish between outliers and non-outliers. Especially for online learning tasks, one is interested in avoiding inappropriate weight changes due to measurement outliers to maintain stable online learning capability. We show experimentally that using the -distribution as a noise model leads to stable online learning algorithms and outperforms state-of-the art online learning methods like the extended Kalman filter algorithm.

 
 


© 2010 The MIT Press
MIT Logo