Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Computational power of neural networks: a characterization in terms of Kolmogorov complexity

Published: 01 September 2006 Publication History

Abstract

The computational power of recurrent neural networks is shown to depend ultimately on the complexity of the real constants (weights) of the network. The complexity, or information contents, of the weights is measured by a variant of resource-bounded Kolmogorov (1965) complexity, taking into account the time required for constructing the numbers. In particular, we reveal a full and proper hierarchy of nonuniform complexity classes associated with networks having weights of increasing Kolmogorov complexity

Cited By

View all
  1. Computational power of neural networks: a characterization in terms of Kolmogorov complexity

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image IEEE Transactions on Information Theory
      IEEE Transactions on Information Theory  Volume 43, Issue 4
      July 1997
      296 pages

      Publisher

      IEEE Press

      Publication History

      Published: 01 September 2006

      Qualifiers

      • Research-article

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)0
      • Downloads (Last 6 weeks)0
      Reflects downloads up to 03 Oct 2024

      Other Metrics

      Citations

      Cited By

      View all

      View Options

      View options

      Get Access

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media