288 pp. per issue
6 x 9, illustrated
2014 Impact factor:

Neural Computation

January 1, 2005, Vol. 17, No. 1, Pages 177-204
(doi: 10.1162/0899766052530802)
© 2004 Massachusetts Institute of Technology
On Learning Vector-Valued Functions
Article PDF (170.19 KB)

In this letter, we provide a study of learning in a Hilbert space of vector-valued functions. We motivate the need for extending learning theory of scalar-valued functions by practical considerations and establish some basic results for learning vector-valued functions that should prove useful in applications. Specifically, we allow an output space Y to be a Hilbert space, and we consider a reproducing kernel Hilbert space of functions whose values lie in Y. In this setting, we derive the form of the minimal norm interpolant to a finite set of data and apply it to study some regularization functionals that are important in learning theory. We consider specific examples of such functionals corresponding to multiple-output regularization networks and support vector machines, for both regression and classification. Finally, we provide classes of operator-valued kernels of the dot product and translation-invariant type.