Nothing Special   »   [go: up one dir, main page]

skip to main content
article
Free access

Large-Sample Learning of Bayesian Networks is NP-Hard

Published: 01 December 2004 Publication History

Abstract

In this paper, we provide new complexity results for algorithms that learn discrete-variable Bayesian networks from data. Our results apply whenever the learning algorithm uses a scoring criterion that favors the simplest structure for which the model is able to represent the generative distribution exactly. Our results therefore hold whenever the learning algorithm uses a consistent scoring criterion and is applied to a sufficiently large dataset. We show that identifying high-scoring structures is NP-hard, even when any combination of one or more of the following hold: the generative distribution is perfect with respect to some DAG containing hidden variables; we are given an independence oracle; we are given an inference oracle; we are given an information oracle; we restrict potential solutions to structures in which each node has at most k parents, for all k>=3.Our proof relies on a new technical result that we establish in the appendices. In particular, we provide a method for constructing the local distributions in a Bayesian network such that the resulting joint distribution is provably perfect with respect to the structure of the network.

References

[1]
Bouckaert, R. R. (1995). Bayesian Belief Networks: From Construction to Inference. PhD thesis, Utrecht University, The Netherlands.
[2]
Chickering, D. M. (1995). A transformational characterization of Bayesian network structures. In Hanks, S. and Besnard, P., editors, Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence, Montreal, QU, pages 87-98. Morgan Kaufmann.
[3]
Chickering, D. M. (1996). Learning Bayesian networks is NP-Complete. In Fisher, D. and Lenz, H., editors, Learning from Data: Artificial Intelligence and Statistics V, pages 121-130. Springer-Verlag.
[4]
Chickering, D. M. (2002). Optimal structure identification with greedy search. Journal of Machine Learning Research, 3:507-554.
[5]
Dasgupta, S. (1999). Learning polytrees. In Laskey, K. and Prade, H., editors, Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence, Stockholm, Sweden, pages 131- 141. Morgan Kaufmann.
[6]
Druzdzel, M. J. and Henrion, M. (1993). Efficient reasoning in qualitative probabilistic networks. In Proceedings of the Eleventh Annual Conference on Artificial Intelligence (AAAI-93), Washington, D.C., pages 548-553.
[7]
Garey, M. and Johnson, D. (1979). Computers and intractability: A guide to the theory of NP-completeness . W.H. Freeman.
[8]
Gavril, F. (1977). Some NP-complete problems on graphs. In Proc. 11th Conf. on Information Sciences and Systems, Johns Hopkins University, pages 91-95. Baltimore, MD.
[9]
Howard, R. and Matheson, J. (1981). Influence diagrams. In Readings on the Principles and Applications of Decision Analysis, volume II, pages 721-762. Strategic Decisions Group, Menlo Park, CA.
[10]
Karlin, S. and Rinott, Y. (1980). Classes of orderings of measures and related correlation inequalities. i. multivariate totally positive distributions. Journal of Multivariate Analysis, 10:467-498.
[11]
Meek, C. (1997). Graphical Models: Selecting causal and statistical models. PhD thesis, Carnegie Mellon University.
[12]
Meek, C. (2001). Finding a path is harder than finding a tree. Journal of Artificial Intelligence Research, 15:383-389.
[13]
Nielsen, J. D., Kočka, T., and Peña, J. M. (2003). On local optima in learning Bayesian networks. In Proceedings of the Nineteenth Conference on Uncertainty in Artificial Intelligence, Acapulco, Mexico, pages 435-442. Morgan Kaufmann.
[14]
Pearl, J. (1988). Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, San Mateo, CA.
[15]
Spirtes, P., Glymour, C., and Scheines, R. (2000). Causation, Prediction, and Search (second edition). The MIT Press, Cambridge, Massachussets.
[16]
Srebro, N. (2001). Maximum likelihood bounded tree-width Markov networks. In Breese, J. and Koller, D., editors, Proceedings of the Seventeenth Conference on Uncertainty in Artificial Intelligence , Seattle, WA, pages 504-511. Morgan Kaufmann.
[17]
Wellman, M. P. (1990). Fundamental concepts of qualitative probabilistic networks. Artificial Intelligence, 44:257-303.

Cited By

View all
  • (2024)Parallel structural learning of Bayesian networksKnowledge-Based Systems10.1016/j.knosys.2024.111840296:COnline publication date: 19-Jul-2024
  • (2024)Finding community structure in Bayesian networks by heuristic K-standard deviation methodFuture Generation Computer Systems10.1016/j.future.2024.03.047158:C(556-568)Online publication date: 1-Sep-2024
  • (2023)Distributionally robust skeleton learning of discrete Bayesian networksProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3668887(63343-63371)Online publication date: 10-Dec-2023
  • Show More Cited By
  1. Large-Sample Learning of Bayesian Networks is NP-Hard

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image The Journal of Machine Learning Research
      The Journal of Machine Learning Research  Volume 5, Issue
      12/1/2004
      1571 pages
      ISSN:1532-4435
      EISSN:1533-7928
      Issue’s Table of Contents

      Publisher

      JMLR.org

      Publication History

      Published: 01 December 2004
      Published in JMLR Volume 5

      Qualifiers

      • Article

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)61
      • Downloads (Last 6 weeks)16
      Reflects downloads up to 25 Nov 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Parallel structural learning of Bayesian networksKnowledge-Based Systems10.1016/j.knosys.2024.111840296:COnline publication date: 19-Jul-2024
      • (2024)Finding community structure in Bayesian networks by heuristic K-standard deviation methodFuture Generation Computer Systems10.1016/j.future.2024.03.047158:C(556-568)Online publication date: 1-Sep-2024
      • (2023)Distributionally robust skeleton learning of discrete Bayesian networksProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3668887(63343-63371)Online publication date: 10-Dec-2023
      • (2023)iSCANProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3668057(44671-44706)Online publication date: 10-Dec-2023
      • (2023)Global optimality in bivariate gradient-based DAG learningProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3666910(17929-17968)Online publication date: 10-Dec-2023
      • (2023)Learning DAGs from data with few root causesProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3666860(16865-16888)Online publication date: 10-Dec-2023
      • (2023)DRCFSProceedings of the 40th International Conference on Machine Learning10.5555/3618408.3619589(28468-28491)Online publication date: 23-Jul-2023
      • (2023)Structural Hawkes processes for learning causal structure from discrete-time event sequencesProceedings of the Thirty-Second International Joint Conference on Artificial Intelligence10.24963/ijcai.2023/633(5702-5710)Online publication date: 19-Aug-2023
      • (2023)A Breast Cancer Detection Method Based on Bayesian NetworksProceedings of the 4th International Conference on Artificial Intelligence and Computer Engineering10.1145/3652628.3652783(933-937)Online publication date: 17-Nov-2023
      • (2023)Feature Selection for Efficient Local-to-global Bayesian Network Structure LearningACM Transactions on Knowledge Discovery from Data10.1145/362447918:2(1-27)Online publication date: 19-Sep-2023
      • Show More Cited By

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Login options

      Full Access

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media