Online learning with kernels. Kivinen, J., Smola, A., & Williamson, R. IEEE Transactions on Signal Processing, 52(8):2165–2176, August, 2004. Conference Name: IEEE Transactions on Signal Processing
doi  abstract   bibtex   
Kernel-based algorithms such as support vector machines have achieved considerable success in various problems in batch setting, where all of the training data is available in advance. Support vector machines combine the so-called kernel trick with the large margin idea. There has been little use of these methods in an online setting suitable for real-time applications. In this paper, we consider online learning in a reproducing kernel Hilbert space. By considering classical stochastic gradient descent within a feature space and the use of some straightforward tricks, we develop simple and computationally efficient algorithms for a wide range of problems such as classification, regression, and novelty detection. In addition to allowing the exploitation of the kernel trick in an online setting, we examine the value of large margins for classification in the online setting with a drifting target. We derive worst-case loss bounds, and moreover, we show the convergence of the hypothesis to the minimizer of the regularized risk functional. We present some experimental results that support the theory as well as illustrating the power of the new algorithms for online novelty detection.
@article{kivinen_online_2004,
	title = {Online learning with kernels},
	volume = {52},
	issn = {1941-0476},
	doi = {10.1109/TSP.2004.830991},
	abstract = {Kernel-based algorithms such as support vector machines have achieved considerable success in various problems in batch setting, where all of the training data is available in advance. Support vector machines combine the so-called kernel trick with the large margin idea. There has been little use of these methods in an online setting suitable for real-time applications. In this paper, we consider online learning in a reproducing kernel Hilbert space. By considering classical stochastic gradient descent within a feature space and the use of some straightforward tricks, we develop simple and computationally efficient algorithms for a wide range of problems such as classification, regression, and novelty detection. In addition to allowing the exploitation of the kernel trick in an online setting, we examine the value of large margins for classification in the online setting with a drifting target. We derive worst-case loss bounds, and moreover, we show the convergence of the hypothesis to the minimizer of the regularized risk functional. We present some experimental results that support the theory as well as illustrating the power of the new algorithms for online novelty detection.},
	number = {8},
	journal = {IEEE Transactions on Signal Processing},
	author = {Kivinen, J. and Smola, A.J. and Williamson, R.C.},
	month = aug,
	year = {2004},
	note = {Conference Name: IEEE Transactions on Signal Processing},
	keywords = {Australia, Condition monitoring, Convergence, Gaussian processes, Hilbert space, Kernel, Signal processing algorithms, Stochastic processes, Support vector machines, Training data},
	pages = {2165--2176},
}

Downloads: 0