Robust C-loss kernel classifiers. Xu, G., Hu, B., & Principe, J. C. IEEE Transactions on Neural Networks and Learning Systems, 29(3):510–522, March, 2018.
Robust C-loss kernel classifiers [link]Paper  doi  abstract   bibtex   
The correntropy-induced loss (C-loss) function has the nice property of being robust to outliers. In this paper, we study the C-loss kernel classifier with the Tikhonov regularization term, which is used to avoid overfitting. After using the halfquadratic optimization algorithm, which converges much faster than the gradient optimization algorithm, we find out that the resulting C-loss kernel classifier is equivalent to an iterative weighted least square support vector machine (LS-SVM). This relationship helps explain the robustness of iterative weighted LS-SVM from the correntropy and density estimation perspectives. On the large-scale data sets which have low-rank Gram matrices, we suggest to use incomplete Cholesky decomposition to speed up the training process. Moreover, we use the representer theorem to improve the sparseness of the resulting C-loss kernel classifier. Experimental results confirm that our methods are more robust to outliers than the existing common classifiers.
@article{xu_robust_2018,
	title = {Robust {C}-loss kernel classifiers},
	volume = {29},
	issn = {2162-237X, 2162-2388},
	url = {https://ieeexplore.ieee.org/document/7801825/},
	doi = {10.1109/TNNLS.2016.2637351},
	abstract = {The correntropy-induced loss (C-loss) function has the nice property of being robust to outliers. In this paper, we study the C-loss kernel classifier with the Tikhonov regularization term, which is used to avoid overfitting. After using the halfquadratic optimization algorithm, which converges much faster than the gradient optimization algorithm, we find out that the resulting C-loss kernel classifier is equivalent to an iterative weighted least square support vector machine (LS-SVM). This relationship helps explain the robustness of iterative weighted LS-SVM from the correntropy and density estimation perspectives. On the large-scale data sets which have low-rank Gram matrices, we suggest to use incomplete Cholesky decomposition to speed up the training process. Moreover, we use the representer theorem to improve the sparseness of the resulting C-loss kernel classifier. Experimental results confirm that our methods are more robust to outliers than the existing common classifiers.},
	language = {en},
	number = {3},
	urldate = {2023-12-01},
	journal = {IEEE Transactions on Neural Networks and Learning Systems},
	author = {Xu, Guibiao and Hu, Bao-Gang and Principe, Jose C.},
	month = mar,
	year = {2018},
	keywords = {/unread},
	pages = {510--522},
}

Downloads: 0