| Article ID | Journal | Published Year | Pages | File Type | 
|---|---|---|---|---|
| 5773528 | Applied and Computational Harmonic Analysis | 2018 | 21 Pages | 
Abstract
												The correntropy-induced loss (C-loss) has been employed in learning algorithms to improve their robustness to non-Gaussian noise and outliers recently. Despite its success on robust learning, only little work has been done to study the generalization performance of regularized regression with the C-loss. To enrich this theme, this paper investigates a kernel-based regression algorithm with the C-loss and â1-regularizer in data dependent hypothesis spaces. The asymptotic learning rate is established for the proposed algorithm in terms of novel error decomposition and capacity-based analysis technique. The sparsity characterization of the derived predictor is studied theoretically. Empirical evaluations demonstrate its advantages over the related approaches.
											Related Topics
												
													Physical Sciences and Engineering
													Mathematics
													Analysis
												
											Authors
												Hong Chen, Yulong Wang, 
											