Abstract
We present a method to learn and select a succinct multi-layer perceptron having shared weights. Weight sharing means a weight is allowed to have one of common weights. A near-zero common weight can be eliminated, called weight pruning. Our method iteratively merges and splits common weights based on 2nd-order criteria, escaping local optima through bidirectional clustering. Moreover, our method selects the optimal number of hidden units based on cross-validation. Our experiments showed that the proposed method can perfectly restore the original sharing structure for an artificial data set, and finds a small number of common weights for a real data set.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Bishop, C.M.: Neural networks for pattern recognition. Clarendon Press (1995)
Haykin, S.: Neural networks, 2nd edn. Prentice-Hall, Englewood Cliffs (1999)
Saito, K., Nakano, R.: Structuring neural networks through bidirectional clustering of weights. In: Proc. 5th Int. Conf. on Discovery Science, pp. 206–219 (2002)
Stone, M.: Cross-validatory choice and assessment of statistical predictions. Journal of the Royal Statistical Society B 64, 111–147 (1974)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Tanahashi, Y., Saito, K., Nakano, R. (2005). Model Selection and Weight Sharing of Multi-layer Perceptrons. In: Khosla, R., Howlett, R.J., Jain, L.C. (eds) Knowledge-Based Intelligent Information and Engineering Systems. KES 2005. Lecture Notes in Computer Science(), vol 3684. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11554028_100
Download citation
DOI: https://doi.org/10.1007/11554028_100
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-28897-8
Online ISBN: 978-3-540-31997-9
eBook Packages: Computer ScienceComputer Science (R0)