Learning rates of multi-kernel regularized regression
From MaRDI portal
Publication:974504
DOI10.1016/j.jspi.2010.03.020zbMath1206.68252MaRDI QIDQ974504
Publication date: 3 June 2010
Published in: Journal of Statistical Planning and Inference (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.jspi.2010.03.020
reproducing kernel Hilbert spaces; learning rate; multi-kernel regularization; Rademacher chaos complexity
Related Items
Refined Rademacher Chaos Complexity Bounds with Applications to the Multikernel Learning Problem, Statistical analysis of the moving least-squares method with unbounded sampling, Approximation analysis of gradient descent algorithm for bipartite ranking, Reproducing kernels and choices of associated feature spaces, in the form of \(L^2\)-spaces, Learning performance of regularized regression with multiscale kernels based on Markov observations, Optimal convergence rates of high order Parzen windows with unbounded sampling
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Model selection for regularized least-squares algorithm in learning theory
- Multi-kernel regularized classifiers
- Learning and approximation by Gaussians on Riemannian manifolds
- Fast rates for support vector machines using Gaussian kernels
- Analysis of support vector machines regression
- Ranking and empirical minimization of \(U\)-statistics
- Learning rates of least-square regularized regression
- On the mathematical foundations of learning
- Learning Theory
- SVM Soft Margin Classifiers: Linear Programming versus Quadratic Programming
- 10.1162/153244303321897690
- Neural Network Learning
- Learning Bounds for Support Vector Machines with Learned Kernels
- Theory of Reproducing Kernels