Nothing Special   »   [go: up one dir, main page]

skip to main content
article

Learning the mean: A neural network approach

Published: 01 February 2012 Publication History

Abstract

One of the key problems in machine learning theory and practice is setting the correct value of the regularization parameter; this is particularly crucial in Kernel Machines such as Support Vector Machines, Regularized Least Square or Neural Networks with Weight Decay terms. Well known methods such as Leave-One-Out (or GCV) and Evidence Maximization offer a way of predicting the regularization parameter. This work points out the failure of these methods for predicting the regularization parameter when coping with the, apparently trivial and here introduced, regularized mean problem; this is the simplest form of Tikhonov regularization, that, in turn, is the primal form of the learning algorithm Regularized Least Squares. This controlled environment gives the possibility to define oracular notions of regularization and to experiment new methodologies for predicting the regularization parameter that can be extended to the more general regression case. The analysis stems from James-Stein theory, shows the equivalence of shrinking and regularization and is carried using multiple kernels learning for regression and SVD analysis; a mean value estimator is built, first via a rational function and secondly via a balanced neural network architecture suitable for estimating statistical quantities and gaining symmetric expectations. The obtained results show that a non-linear analysis of the sample and a non-linear estimation of the mean obtained by neural networks can be profitably used to improve the accuracy of mean value estimations, especially when a small number of realizations is provided.

References

[1]
Vapnik, V., Statistical Learning Theory. 1998. Wiley-Interscience Pub.
[2]
Evgeniou, T., Pontil, M. and Poggio, T., Regularization networks and support vector machines. Advances in Computational Mathematics.
[3]
W. James, C. Stein, Estimation with quadratic loss, in: Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, vol. 1, 1961, pp. 361-379.
[4]
Rakotomamonjy, A., Bach, F.R., Canu, S. and GrandValet, Y., SimpleMKL. Journal of Machine Learning Research. 2491-2521.
[5]
Moody, J.E., The effective number of parameters: an analysis of generalisation and regularisation in nonlinear learning systems. In: Moody, J.E., Hanson, S.J., Lippmann, R.P. (Eds.), Neural Information Processing Systems, 4. Morgan Kaufmann, San Mateo CA. pp. 847-854.
[6]
Geisser, S., Predictive Inference. 1993. Chapman and Hall, New York.
[7]
MacKay, D.J.C., A practical Bayesian framework for backpropagation networks. Neural Computation. v4 i3. 448-472.
[8]
Ridella, S., Rovetta, S. and Zunino, R., Circular backpropagation networks for classification. IEEE Transactions On Neural Networks. v8. 84-97.
[9]
Thompson, J.R., Some shrinkage techniques for estimating the mean. Journal of the American Statistical Association. v63 i321. 113-122.
[10]
Borgwardt, K., Gretton, A., Rasch, M., Kriegel, H.-P., Schölkopf, B. and Smola, A., Integrating structured biological data by kernel maximum mean discrepancy. Bioinformatics. v22 i4. 49-57.
[11]
Liang, P., Bach, F., Bouchard, G. and Jordan, M.I., Asymptotically optimal regularization in smooth parametric models. Advances in Neural Information Processing Systems (NIPS).
[12]
Mandelkern, M., Setting confidence intervals for bounded parameters. Statistical Science. v17. 149
[13]
S. Decherchi, M. Parodi, S. Ridella, A neural model approach for regularization in the man estimation case, in: International Joint Conference on Neural Networks, Barcellona, 2010.

Cited By

View all
  • (2018)Currency Exchange Rate Forecasting Using Artificial Neural Networks Backpropagation MethodInternational Journal of Green Computing10.4018/jgc.20120701023:2(14-33)Online publication date: 11-Dec-2018
  1. Learning the mean: A neural network approach

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image Neurocomputing
    Neurocomputing  Volume 77, Issue 1
    February, 2012
    289 pages

    Publisher

    Elsevier Science Publishers B. V.

    Netherlands

    Publication History

    Published: 01 February 2012

    Author Tags

    1. Back-propagation
    2. James-Stein theory
    3. Mean problem
    4. Model selection
    5. Multiple kernel learning
    6. Regularization
    7. SVD
    8. Shrinkage

    Qualifiers

    • Article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 18 Dec 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2018)Currency Exchange Rate Forecasting Using Artificial Neural Networks Backpropagation MethodInternational Journal of Green Computing10.4018/jgc.20120701023:2(14-33)Online publication date: 11-Dec-2018

    View Options

    View options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media