Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/2505515.2505584acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

Eigenvalues perturbation of integral operator for kernel selection

Published: 27 October 2013 Publication History

Abstract

Kernel selection is one of the key issues both in recent research and application of kernel methods. This is usually done by minimizing either an estimate of generalization error or some other related performance measure. It is well known that a kernel matrix can be interpreted as an empirical version of a continuous integral operator, and its eigenvalues converge to the eigenvalues of integral operator. In this paper, we introduce new kernel selection criteria based on the eigenvalues perturbation of the integral operator. This perturbation quantifies the difference between the eigenvalues of the kernel matrix and those of the integral operator. We establish the connection between eigenvalues perturbation and generalization error. By minimizing the derived generalization error bounds, we propose the kernel selection criteria. Therefore the kernel chosen by our proposed criteria can guarantee good generalization performance. To compute the values of our criteria, we present a method to obtain the eigenvalues of integral operator via the Fourier transform. Experiments on benchmark datasets demonstrate that our kernel selection criteria are sound and effective.

References

[1]
F. Bach, G. Lanckriet, and M. Jordan. Multiple kernel learning, conic duality, and the SMO algorithm. In Proceedings of the 21th International Conference on Machine Learning (ICML), pages 41--48, 2004.
[2]
P. L. Bartlett, S. Boucheron, and G. Lugosi. Model selection and error estimation. Machine Learning, 48:85--113, 2002.
[3]
P. L. Bartlett and S. Mendelson. Rademacher and gaussian complexities: Risk bounds and structural results. The Journal of Machine Learning Research, 3:463--482, 2002.
[4]
O. Bousquet and A. Elisseeff. Stability and generalization. Journal of Machine Learning Research, 2:499--526, 2002.
[5]
M. Braun. Accurate error bounds for the eigenvalues of the kernel matrix. The Journal of Machine Learning Research, 7:2303--2328, 2006.
[6]
G. C. Cawley and N. L. C. Talbot. Preventing over-fitting during model selection via bayesian regularisation of the hyper-parameters. Journal of Machine Learning Research, 8:841--861, 2007.
[7]
G. C. Cawley and N. L. C. Talbot. On over-fitting in model selection and subsequent selection bias in performance evaluation. Journal of Machine Learning Research, 11:2079--2107, 2010.
[8]
O. Chapelle and V. Vapnik. Model selection for Support Vector Machines. In Advances in Neural Information Processing Systems 12 (NIPS), pages 230--236, 1999.
[9]
O. Chapelle, V. Vapnik, O. Bousquet, and S. Mukherjee. Choosing multiple parameters for support vector machines. Machine Learning, 46(1-3):131--159, 2002.
[10]
C. Corte and V. Vapnik. Support-vector networks. Machine Learning, 20(3):273--297, 1995.
[11]
C. Cortes, M. Mohri, and A. Rostamizadeh. L2 regularization for learning kernels. In Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence (UAI), pages 109--116, 2009.
[12]
C. Cortes, M. Mohri, and A. Rostamizadeh. Two-stage learning kernel algorithms. In Proceedings of the 27th Conference on Machine Learning (ICML), pages 239--246, 2010.
[13]
C. Cortes, M. Mohri, and A. Talwalkar. On the impact of kernel approximation on learning accuracy. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics (AISTATS), pages 113--120, 2000.
[14]
N. Cristianini, J. Shawe-Taylor, A. Elisseeff, and J. S. Kandola. On kernel-target alignment. In Advances in Neural Information Processing Systems 14 (NIPS), pages 367--373, 2001.
[15]
M. Debruyne, M. Hubert, and J. A. Suykens. Model selection in kernel based regression using the influence function. Journal of Machine Learning Research, 9:2377--2400, 2008.
[16]
T. Evgeniou, M. Pontil, and T. Poggio. Regularization networks and Support Vector Machines. Advances in Computational Mathematics, 13:1--50, 2000.
[17]
C. S. A. Gammerman and V. Vovk. Ridge regression learning algorithm in dual variables. In Proceedings of the 15th International Conference on Machine Learning (ICML), pages 515--521, 1998.
[18]
L. L. Gerfo, L. Rosasco, F. Odone, E. D. Vito, and A. Verri. Spectral algorithms for supervised learning. Neural Computation, 20(3):331--368, 2008.
[19]
G. H. Golub, M. Heath, and G. Wahba. Generalized cross-validation as a method for choosing a good ridge parameter. Technometrics, 21(2):215--223, 1979.
[20]
L. Jia and S. Liao. Accurate probabilistic error bound for eigenvalues of kernel matrix. In Proceedings of the 1st Asian Conference on Machine Learning,2009.
[21]
M. Kloft, U. Brefeld, S. Sonnenburg, and A. Zien. Lp-norm multiple kernel learning. Journal of Machine Learning Research, 12:953--997, 2011.
[22]
G. R. G. Lanckriet, N. Cristianini, P. L. Bartlett, L. E. Ghaoui, and M. I. Jordan. Learning the kernel matrix with semidefinite programming. The Journal of Machine Learning Research, 5:27--72, 2004.
[23]
Y. Liu, S. Liao, and Y. Hou. Learning kernels with upper bounds of leave-one-out error. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management (CIKM), pages 2205--2208, 2011.
[24]
U. V. Luxburg, O. Bousquet, and B. Schölkopf. A compression approach to Support Vector model selection. Journal of Machine Learning Research, 5:293--323, 2004.
[25]
C. H. Nguyen and T. B. Ho. Kernel matrix evaluation. In Proceedings of the 20th International Joint Conference on Artifficial Intelligence (IJCAI), pages 987--992, 2007.
[26]
C. H. Nguyen and T. B. Ho. An efficient kernel matrix evaluation measure. Pattern Recognition, 41(11):3366--3372, 2008.
[27]
A. Rakotomamonjy, F. Bach, S. Canu, and Y. Grandvalet. SimpleMKL. Journal of Machine Learning Research, 9:2491--2521, 2008.
[28]
L. Rosasco, M. Belkin, and E. Vito. On learning with integral operators. The Journal of Machine Learning Research, 11:905--934, 2010.
[29]
B. Schölkopf and A. J. Smola. Learning with kernels. MIT Press, Cambridge, MA, 2002.
[30]
J. Shawe-Taylor and N. Cristianini. Support Vector Machines. Cambridge University Press, MA, 2000.
[31]
I. Steinwart and A. Christmann. Support Vector Machines. Springer Verlag, 2008.
[32]
J. A. K. Suykens and J. Vandewalle. Least squares Support Vector Machine classifiers. Neural Processing Letters, 9(3):293--300, 1999.
[33]
V. Vapnik. The Nature of Statistical Learning Theory. Springer Verlag, 2000.
[34]
G. Wahba. Support Vector Machine, reproducing kernel Hilbert spaces and the randomized GACV. In Advances in Kernel Methods-Support Vector Learning, volume 6, pages 69--88, Cambridge, 1999. MIT Press.
[35]
G. Wahba, Y. Lin, and H. Zhang. GACV for support vector machines. In Advances in Large Margin Classifiers. MIT Press, Cambridge, 1999.

Cited By

View all
  • (2022)Preventing Over-Fitting of Cross-Validation with Kernel StabilityMachine Learning and Knowledge Discovery in Databases10.1007/978-3-662-44851-9_19(290-305)Online publication date: 10-Mar-2022
  • (2021)Kernel Stability for Model Selection in Kernel-Based AlgorithmsIEEE Transactions on Cybernetics10.1109/TCYB.2019.292382451:12(5647-5658)Online publication date: Dec-2021
  • (2018)Multi-class learningProceedings of the 32nd International Conference on Neural Information Processing Systems10.5555/3326943.3327089(1593-1602)Online publication date: 3-Dec-2018
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
CIKM '13: Proceedings of the 22nd ACM international conference on Information & Knowledge Management
October 2013
2612 pages
ISBN:9781450322638
DOI:10.1145/2505515
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 27 October 2013

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. eigenvalues perturbation
  2. generalization error
  3. integral operator
  4. kernel selection

Qualifiers

  • Research-article

Conference

CIKM'13
Sponsor:
CIKM'13: 22nd ACM International Conference on Information and Knowledge Management
October 27 - November 1, 2013
California, San Francisco, USA

Acceptance Rates

CIKM '13 Paper Acceptance Rate 143 of 848 submissions, 17%;
Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

Upcoming Conference

CIKM '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)6
  • Downloads (Last 6 weeks)3
Reflects downloads up to 18 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2022)Preventing Over-Fitting of Cross-Validation with Kernel StabilityMachine Learning and Knowledge Discovery in Databases10.1007/978-3-662-44851-9_19(290-305)Online publication date: 10-Mar-2022
  • (2021)Kernel Stability for Model Selection in Kernel-Based AlgorithmsIEEE Transactions on Cybernetics10.1109/TCYB.2019.292382451:12(5647-5658)Online publication date: Dec-2021
  • (2018)Multi-class learningProceedings of the 32nd International Conference on Neural Information Processing Systems10.5555/3326943.3327089(1593-1602)Online publication date: 3-Dec-2018
  • (2018)Fast cross-validationProceedings of the 27th International Joint Conference on Artificial Intelligence10.5555/3304889.3305007(2497-2503)Online publication date: 13-Jul-2018
  • (2017)Efficient kernel selection via spectral analysisProceedings of the 26th International Joint Conference on Artificial Intelligence10.5555/3172077.3172183(2124-2130)Online publication date: 19-Aug-2017
  • (2017)Granularity selection for cross-validation of SVMInformation Sciences: an International Journal10.1016/j.ins.2016.06.051378:C(475-483)Online publication date: 1-Feb-2017
  • (2014)Model Selection with the Covering Number of the Ball of RKHSProceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management10.1145/2661829.2662034(1159-1168)Online publication date: 3-Nov-2014

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media