Abstract
A crucial characteristic of machine learning models in various domains (such as medical diagnosis, financial analysis, or real-time process monitoring) is the interpretability. The interpretation supports humans in understanding the meaning behind every single prediction made by the machine, and enables the user to assess trustworthiness before acting on the predictions. This article presents our work in building an interpretable classification model based on association rule mining and multi-objective optimization. The classification model itself is a rule list, making a single prediction based on multiple rules. The rule list consists of If ... THEN statements that are understandable to humans. We choose these rules from a large set of pre-mined rules according to an interestingness measure which is formulated as a function of basic probabilities related to the rules. We learned the interestingness measure through multi-objective optimization, concentrating on two objectives: the classifier’s size in terms of number of rules and prediction accuracy. The model is called MoMAC, “Multi-Objective optimization to combine Multiple Association rules into an interpretable Classification”. The experimental results on benchmark datasets demonstrate that MoMAC outperforms other existing rule-based classification methods in terms of classification accuracy.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Code Availability
References
Agrawal R, Srikant R (1994) Fast algorithms for mining association rules. Proceedings of the Interntational Conference on Very Large Databases: 487–499
Angelino E, Larus-Stone N, Alabi D, Seltzer M, Rudin C (2017) Learning certifiably optimal rule lists for categorical data. J Mach Learn Res 18:8753–8830
Bertsimas D, Dunn J (2017) Optimal classification trees. Mach Learn 106:1039–1082
Chen C, Rudin C (2018) An optimization approach to learning falling rule lists. AISTATS: 604–612
Dash P, Fiore-Gartland AJ, Hertz T, Wang GC, Sharma S, Souquette A, Crawford JC, Clemens EB, Nguyen TH, Kedzierska K, et al. (2017) Quantifiable predictive features define epitope-specific t cell receptor repertoires. Nature 547:89–93
Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multi-objective genetic algorithm: NSGA-II. IEEE Trans Evol Comput 6:182–197
Dembczyński K., Kotłowski W, Słowiński R (2008) Maximum likelihood rule ensembles. Proceedings of the International Conference on Machine Learning: 224–231
Dembczyński K., Kotłowski W, Słowiński R (2010) Ender: a statistical framework for boosting decision rules. Data Mining and Knowledge Discovery 21:52–90
Djenouri Y, Belhadi A, Fournier-Viger P, Fujita H (2018) Mining diversified association rules in big datasets: A cluster/GPU/genetic approach. Inform Sci 459:117–134
Djenouri Y, Lin JCW, Nørvåg K, Ramampiaro H (2019) Highly efficient pattern mining based on transaction decomposition. In: 2019 IEEE 35th International Conference on Data Engineering (ICDE). IEEE, pp 1646–1649
Dong G, Li J (2009) Emerging pattern based classification. Encyclopedia of Database Systems: 985
Dong G, Zhang X, Wong L, Li J (1999) CAEP: Classification by aggregating emerging patterns. International Conference on Discovery Science: 30–42
Dua D, Graff C (2019) UCI machine learning repository. http://archive.ics.uci.edu/ml1646–1649
Fan H, Ramamohanarao K (2006) Fast discovery and the generalization of strong jumping emerging patterns for building compact and accurate classifiers. IEEE Trans Knowl Data Eng 18:721–737
Friedman JH, Popescu BE (2008) Predictive learning via rule ensembles. Annals Appl Stat 2:916–954
Geng L, Hamilton HJ (2006) Interestingness measures for data mining: A survey. ACM Computing Surveys (CSUR) 38:9–es
Lakkaraju H, Bach SH, Leskovec J (2016) Interpretable decision sets: A joint framework for description and prediction. Proceedings of the International Conference on Knowledge Discovery and Data Mining: 1675–1684
Letham B, Rudin C, McCormick TH, Madigan D (2015) Interpretable classifiers using rules and bayesian analysis: Building a better stroke prediction model. Annals Appl Stat 9:1350–1371
Li J, Dong G, Ramamohanarao K (2000) Instance-based classification by emerging patterns. European Conference on Principles of Data Mining and Knowledge Discovery: 191– 200
Li W, Han J, Pei J (2001) CMAR: accurate and efficient classification based on multiple class-association rules. IEEE International Conference on Data Mining: 369–376
Liu B, Hsu W, Ma Y (1998) Integrating classification and association rule mining. KDD: 80–86
Luna JM, Fournier-Viger P, Ventura S (2019) Frequent itemset mining: A 25 years review. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 9(6)
Mattiev J, Kavšek B (2020) A compact and understandable associative classifier based on overall coverage. Procedia Computer Science 170:1161–1167
Mita G, Papotti P, Filippone M, Michiardi P (2020) LIBRE: Learning interpretable boolean rule ensembles. In: International conference on artificial intelligence and statistics. PMLR, pp 245–255
Molnar C (2018) Interpretable machine learning: a guide for making black box models explainable. Leanpub
New York Civil Liberties Union (2014) Stop and frisk data http://www.nyclu.org/content/stop-and-frisk-data
Paul Y, Kumar N (2020) A comparative study of famous classification techniques and data mining tools. In: Proceedings of ICRIC. Springer, pp 627–644
Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Prettenhofer P, Weiss R, Dubourg V, Vanderplas J, Passos A, Cournapeau D, Brucher M, Perrot M, Duchesnay E (2011) Scikit-learn: Machine learning in python. J Mach Learn Res 12:2825–2830
Quinlan JR (2014) C4.5: Programs for machine learning. Elsevier
Rajab KD (2019) New associative classification method based on rule pruning for classification of datasets. IEEE Access 7 :157783–157795
Rijnbeek PR, Kors JA (2010) Finding a short and accurate decision rule in disjunctive normal form by exhaustive search. Mach Learn 80:33–62
Rudin C (2019) Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat Mach Intell 1:206–215
Sharma R, Kaushik M, Peious SA, Yahia SB, Draheim D (2020) Expected vs. unexpected: selecting right measures of interestingness. In: International conference on big data analytics and knowledge discovery. Springer, pp 38–47
Song K, Lee K (2017) Predictability-based collective class association rule mining. Expert Syst Appl: 1–7
Sood N, Zaiane O (2020) Building a competitive associative classifier. In: International conference on big data analytics and knowledge discovery. Springer, pp 223–234
Venturini L, Baralis E, Garza P (2017) Scaling associative classification for very large datasets. Journal of Big Data 4(1):1–24
Wang F, Rudin C (2015) Falling rule lists. Artif Intell Stat: 1013–1022
Wei D, Dash S, Gao T, Günlük O (2019) Generalized linear rule models. In: Proceedings of the International Conference on Machine Learning, pp 6687–6696
Yang G, Mabu SM, Shimada K, Gong Y, Hirasawa K (2009) Ranking association rules for classification based on genetic network programming. Proceedings of the annual conference on genetic and evolutionary computation: 1917–1918
Yang G, Shimada K, Mabu S, Hirasawa K (2008) A nonlinear model to rank association rules based on semantic similarity and genetic network programming. IEEE Trans Electrical Electronic Eng: 1–9
Yang H, Rudin C, Seltzer M (2017) Scalable bayesian rule lists. Proceedings of the International Conference on Machine Learning 70:3921–3930
Yin X, Han J (2003) CPAR: Classification based on predictive association rules. In: Proceedings of the SIAM international conference on data mining, pp 331–335
Funding
The work was supported by Universiteit Antwerpen under BOF docpro grant to the first author.
Author information
Authors and Affiliations
Corresponding authors
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Bui-Thi, D., Meysman, P. & Laukens, K. MoMAC: Multi-objective optimization to combine multiple association rules into an interpretable classification. Appl Intell 52, 3090–3102 (2022). https://doi.org/10.1007/s10489-021-02595-w
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-02595-w