Nothing Special   »   [go: up one dir, main page]

Skip to main content

Advertisement

Log in

Online Variational Learning of Dirichlet Process Mixtures of Scaled Dirichlet Distributions

  • Published:
Information Systems Frontiers Aims and scope Submit manuscript

Abstract

Data clustering as an unsupervised method has been one of the main attention-grabbing techniques and a large class of tasks can be formulated by this method. Mixture models as a branch of clustering methods have been used in various fields of research such as computer vision and pattern recognition. To apply these models, we need to address some problems such as finding a proper distribution that properly fits data, defining model complexity and estimating the model parameters. In this paper, we apply scaled Dirichlet distribution to tackle the first challenge and propose a novel online variational method to mitigate the other two issues simultaneously. The effectiveness of the proposed work is evaluated by four challenging real applications, namely, text and image spam categorization, diabetes and hepatitis detection.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  • Attias, H. (2000). A variational baysian framework for graphical models. In Advances in neural information processing systems (pp. 209–215).

  • Bdiri, T., & Bouguila, N. (2011). An infinite mixture of inverted dirichlet distributions. In International Conference on Neural Information Processing (pp. 71–78). New York: Springer.

  • Bdiri, T., & Bouguila, B. (2012). Positive vectors clustering using inverted dirichlet finite mixture models. Expert Systems with Applications, 39(2), 1869–1882.

    Google Scholar 

  • Blanzieri, E., & Bryl, A. (2008). A survey of learning-based techniques of email spam filtering. Artificial Intelligence Review, 29(1), 63–92.

    Google Scholar 

  • Blei, D.M., & Jordan, M.I. (2004). Variational methods for the dirichlet process. In proceedings of the twenty-first international conference on Machine learning (pp. 12–20). ACM.

  • Blei, D.M., Jordan, M.I., & et al. (2006). Variational inference for dirichlet process mixtures. Bayesian Analysis, 1(1), 121–143.

    Google Scholar 

  • Bouguila, N. (2012). Infinite liouville mixture models with application to text and texture categorization. Pattern Recognition Letters, 33(2), 103–110.

    Google Scholar 

  • Bouguila, N., & Ziou, D. (2004). Dirichlet-based probability model applied to human skin detection [image skin detection]. In IEEE International Conference on Acoustics, Speech, and Signal Processing, (Vol. 5 pp V–521). IEEE, 2004.

  • Bouguila, N., & Ziou, D. (2006). Unsupervised selection of a finite dirichlet mixture model: an mml-based approach. IEEE Transactions on Knowledge and Data Engineering, 18(8), 993–1009.

    Google Scholar 

  • Bouguila, N., & Ziou, D. (2007). High-dimensional unsupervised selection and estimation of a finite generalized dirichlet mixture model based on minimum message length. IEEE transactions on pattern analysis and machine intelligence, 29(10), 1716–1731.

    Google Scholar 

  • Bouguila, N., & Ziou, D. (2008). A dirichlet process mixture of dirichlet distributions for classification and prediction. In IEEE workshop on machine learning for signal processing, IEEE, 2008 (pp. 297–302).

  • Bouguila, N., & Ziou, D. (2012). A countably infinite mixture model for clustering and feature selection. Knowledge Inform System, 33(2), 351–370.

    Google Scholar 

  • Bishop, C.M. (2006). Pattern recognition and machine learning. Berlin: Springer.

    Google Scholar 

  • Dredze, M., Gevaryahu, R., & Elias-Bachrach, A. (2007). Learning fast classifiers for image spam. In CEAS (pp. 2007–487).

  • Fan, W, & Bouguila, N. (2011). Infinite dirichlet mixture model and its application via variational bayes. In 2011 10th International Conference on Machine Learning and Applications and Workshops, IEEE, (Vol. 1 pp. 129–132).

  • Fan, W., & Bouguila, N. (2012). Online variational learning of finite dirichlet mixture models. Evolving Systems, 3(3), 153–165.

    Google Scholar 

  • Fan, W., & Bouguila, N. (2012). Online learning of a dirichlet process mixture of generalized dirichlet distributions for simultaneous clustering and localized feature selection. In Asian Conference on Machine Learning (pp. 113–128).

  • Fan, W., & Bouguila, N. (2013). Learning finite beta-liouville mixture models via variational bayes for proportional data clustering. In Twenty-Third International Joint Conference on Artificial Intelligence (pp. 1323–1329).

  • Fan, W., & Bouguila, N. (2014). Online variational learning of generalized dirichlet mixture models with feature selection. Neurocomputing, 126, 166–179.

    Google Scholar 

  • Hepatitis. (2013). available at https://www.who.int/features/qa/76/en/e.

  • Hepatitis. (2020). available at https://archive.ics.uci.edu/.

  • Hoffman, M., Bach, F.R., & Blei, D.M. (2010). Online learning for latent dirichlet allocation. In Advances in neural information processing systems (pp. 856–864).

  • Ishwaran, H., & James, L.F. (2001). Gibbs sampling methods for stick-breaking priors. Journal of the American Statistical Association, 96(453), 161–173.

    Google Scholar 

  • Jain, A.K., Murty, M.N., & Flynn, P.J. (1999). Data clustering: a review. ACM computing surveys (CSUR), 31(3), 264–323.

    Google Scholar 

  • Jordan, M.I., Ghahramani, Z., Jaakkola, T.S., & Saul, L.K. (1999). An introduction to variational methods for graphical models. Machine learning, 37(2), 183–233.

    Google Scholar 

  • Kaggle. (1990). diabetes disease dataset. available at https://www.kaggle.com/uciml/pima-indians-diabetes-database.

  • Kaufman, L., & Rousseeuw, P.J. (2009). Finding groups in data: an introduction to cluster analysis Vol. 344. New York: Wiley.

    Google Scholar 

  • Kavakiotis, I., Tsave, O., Salifoglou, A., Maglaveras, N., Vlahavas, I., & Chouvarda, I. (2017). Machine learning and data mining methods in diabetes research. Computational and Structural Biotechnology Journal, 15, 104–116.

    Google Scholar 

  • Kushner, H., & Yin, G.G. (2003). Stochastic approximation and recursive algorithms and applications, volume 35 Springer Science & Business Media.

  • Lowe, D.G. (2004). Distinctive image features from scale-invariant keypoints. International journal of computer vision, 60(2), 91–110.

    Google Scholar 

  • McLachlan, G.J. (2015). Mixture models in statistics.

  • McLachlan, G.J., & Krishnan, T. (2007). The EM algorithm and extensions Vol. 382. New York: Wiley.

    Google Scholar 

  • McLachlan, G., & Peel, D. (2004). Finite mixture models. John Wiley & Sons: New York.

    Google Scholar 

  • Nguyen, H., Kalra, M., Azam, M., & Bouguila, N. (2019). Data clustering using online variational learning of finite scaled dirichlet mixture models. In 20th IEEE International Conference on Information Reuse and Integration for Data Science, IRI 2019 (pp. 267–274).

  • Opper, M, & Saad, D. (2001). Advanced mean field methods: theory and practice neural information processing series.

  • Rasmussen, C.E. (1996). A practical monte carlo implementation of bayesian learning. In Advances in Neural Information Processing Systems (pp. 598–604).

  • Rasmussen, C.E. (2000). The infinite gaussian mixture model. In Advances in neural information processing systems (pp. 554–560).

  • Rasmussen, C.E., & Ghahramani, Z. (2002). Infinite mixtures of gaussian process experts. In Advances in neural information processing systems (pp. 881–888).

  • Sato, M.A. (2001). Online model selection based on the variational bayes. Neural Computation, 13(7), 1649–1681.

    Google Scholar 

  • Sethuraman, J. (1994). A constructive definition of dirichlet priors. Statistica Sinica. 639–650.

  • Spamarchive dataset. (2020). available at https://www.cs.jhu.edu/~mdredze/datasets/image_spam/.

  • Spambase. (2015). available at https://archive.ics.uci.edu/ml/datasets/spambase.

  • WHO diabetes fact sheet. (2018). available at https://www.who.int/news-room/fact-sheets/detail/diabetes/.

  • Zhu, Y., & Tan, Y. (2010). A local-concentration-based feature extraction approach for spam filtering. IEEE Transactions on Information Forensics and Security, 6(2), 486–497.

    Google Scholar 

Download references

Acknowledgements

The completion of this research was made possible thanks to the Natural Sciences and Engineering Research Council of Canada (NSERC) and the National Natural Science Foundation of China (61876068).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Narges Manouchehri.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Manouchehri, N., Nguyen, H., Koochemeshkian, P. et al. Online Variational Learning of Dirichlet Process Mixtures of Scaled Dirichlet Distributions. Inf Syst Front 22, 1085–1093 (2020). https://doi.org/10.1007/s10796-020-10027-2

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10796-020-10027-2

Keywords

Navigation