Abstract
Modern apps based businesses are increasingly interested in data driven decision making to achieve business goals as well as retaining their customer base. In this paper, we propose a Bayesian network approach to assess the mobile apps user behavior. We propose a strategy to build Bayesian networks and further improve the causal networks using re-sampling methods to best represent the causal representation between app user retention and in-app features. Structural hamming distances (SHD) are then used for assessing similar Bayesian network structures learned using data available from a local mobile app developing company. We also conduct a simulation study to assess the effect of re-sampling techniques towards the Bayesian network performance with various learning algorithms.
Similar content being viewed by others
References
Acid S, de Campos LM (2003) Searching for Bayesian network structures in the space of restricted acyclic partially directed graphs. J Artif Int Res 18(1):445–490
Akaike H (1974) A new look at the statistical model identification. IEEE Trans Autom Control 19(6):716–723
Ashcroft M (2012) Bayesian networks in business analytics. In: Proceedings of the federated conference on computer science and information systems, pp 955–961
Barua S, Islam MM, Yao X, Murase K (2014) MWMOTE—majority weighted minority oversampling technique for imbalanced data set learning. IEEE Trans Knowl Data Eng 26(2):405–425. https://doi.org/10.1109/TKDE.2012.232
Bobek S, Baran M, Kluza K, Nalepa GJ (2013) Application of Bayesian networks to recommendations in business process modeling. In: AIBP@ AI* IA, pp 41–50
Bozdogan H (1987) Model selection and Akaike’s information criterion (aic): the general theory and its analytical extensions. Psychometrika 52(3):345–370
Brown LE, Tsamardinos I, Aliferis CF (2004) A novel algorithm for scalable and accurate Bayesian network learning. In: Medinfo, pp 711–715
de Campos LM (2006) A scoring function for learning Bayesian networks based on mutual information and conditional independence tests. J Mach Learn Res 7:2149–2187
Chakraborty S, Mengersen K, Fidge C, Ma L, Lassen D (2016) A Bayesian network-based customer satisfaction model: a tool for management decisions in railway transport. Decis Anal 3(1):4
Chawla N, Japkowicz N, Kołcz A (2004) Editorial: special issue on learning from imbalanced data sets. SIGKDD Explor 6:1–6. https://doi.org/10.1145/1007730.1007733
Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP (2002) SMOTE: synthetic minority over-sampling technique. Tech. rep
Chickering D, Geiger D, Heckerman D (1995) Learning Bayesian networks: search methods and experimental results. In: Preliminary papers of the fifth international workshop on artificial intelligence and statistics
Chickering DM (1996) Learning Bayesian Networks is NP-Complete, learning from data: artificial intelligence and statistics v edn. Springer-Verlag. https://www.microsoft.com/en-us/research/publication/learning-bayesian-networks-is-np-complete/
de Jongh M, Druzdzel MJ (2009) A comparison of structural distance measures for causal bayesian network models. Recent advances in intelligent information systems, challenging problems of science, computer science series, pp 443–456
Dumais S, Platt J, Heckerman D, Sahami M (1998) Inductive learning algorithms and representations for text categorization. In: Proceedings of the seventh international conference on information and knowledge management, CIKM ’98. Association for Computing Machinery, New York, pp 148–155. https://doi.org/10.1145/288627.288651
Ezawa K, Singh M, Norton SW (1996) Learning goal oriented Bayesian networks for telecommunications risk management. In: Proceedings of the 13th international conference on machine learning. Morgan Kaufmann, pp 139–147
Fawcett T, Provost F (1996) Combining data mining and machine learning for effective user profiling. In: Proceedings of the second international conference on knowledge discovery and data mining, KDD’96. AAAI Press, pp 8–13
Fernández A, García S, Herrera F, Chawla NV (2018) SMOTE for learning from imbalanced data: progress and challenges, marking the 15-year anniversary. https://doi.org/10.1613/jair.1.11192
Haibo H, Yang B, Garcia EA, Shutao L (2008) ADASYN: adaptive synthetic sampling approach for imbalanced learning. In: 2008 IEEE international joint conference on neural networks (IEEE world congress on computational intelligence). IEEE, pp 1322–1328. https://doi.org/10.1109/IJCNN.2008.4633969. http://ieeexplore.ieee.org/document/4633969/
Han H, Wang WY, Mao BH (2005) Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. Tech. rep
Hart P (1968) The condensed nearest neighbor rule (corresp.). IEEE Trans Inf Theory 14(3):515–516
Heckerman D, Geiger D, Chickering D (1995) Learning Bayesian networks: the combination of knowledge and statistical data. Mach Learn 20:197–243. https://doi.org/10.1007/BF00994016
Hibbert D, Armstrong N (2009) An introduction to Bayesian methods for analyzing chemistry data: part II: a review of applications of Bayesian methods in chemistry. Chemometr Intell Lab Syst 97(2):211–220
Hoch D (2014) App retention improves—apps used only once declines to 20%. http://info.localytics.com/blog/app-retention-improves
Kubat M, Holte RC, Matwin S (1998) Machine learning for the detection of oil spills in satellite radar images. Mach Learn 30(2):195–215. https://doi.org/10.1023/A:1007452223027
Kubat M, Matwin S (1997) Addressing the curse of imbalanced training sets: one-sided selection. In: ICML
Larranaga P, Poza M, Yurramendi Y, Murga RH, Kuijpers CMH (1996) Structure learning of Bayesian networks by genetic algorithms: a performance analysis of control parameters. IEEE Trans Pattern Anal Mach Intell 18(9):912–926
Laurikkala J (2001) Improving identification of difficult small classes by balancing class distribution. In: Proceedings of the 8th conference on AI in medicine in Europe: artificial intelligence medicine, AIME ’01. Springer, Berlin, pp 63–66
Lewis DD, Catlett J (1994) Heterogeneous uncertainty sampling for supervised learning. In: Cohen WW, Hirsh H (eds) Machine learning proceedings 1994. Morgan Kaufmann, San Francisco, pp 148–156. https://doi.org/10.1016/B978-1-55860-335-6.50026-X
Lewis DD, Ringuette M (1994) A comparison of two learning algorithms for text categorization
Lucas P (2001) Bayesian networks in medicine: a model-based approach to medical decision making
Margaritis D (2003) Learning Bayesian network model structure from data. Carnegie-Mellon Univ Pittsburgh Pa School of Computer Science, Tech. rep
Neapolitan R (2003) Learning Bayesian networks. Prentice Hall. https://doi.org/10.1145/1327942.1327961
Neapolitan RE et al (2004) Learning Bayesian networks, vol 38. Pearson Prentice Hall, Upper Saddle River
Needham CJ, Bradford JR, Bulpitt AJ, Westhead DR (2007) A primer on learning in Bayesian networks for computational biology. PLoS Comput Biol 3(8):1–8. https://doi.org/10.1371/journal.pcbi.0030129
Parsons S (2011) Probabilistic graphical models: principles and techniques. Knowl Eng Rev 26(2):237–238
Pearl J (1988) Probabilistic reasoning in intelligent systems. In: Probabilistic reasoning in intelligent systems. Morgan Kaufmann. https://doi.org/10.1016/B978-0-08-051489-5.50001-1
Perrier E, Imoto S, Miyano S (2008) Finding optimal Bayesian network given a super-structure. J Mach Learn Res 9(Oct):2251–2286
Perro J (2018) Mobile apps: what’s a good retention rate? http://info.localytics.com/blog/mobile-apps-whats-a-good-retention-rate
Provost F, Fawcett T (2001) Robust classification for imprecise environments. Mach Learn 42(3):203–231. https://doi.org/10.1023/A:1007601015854
Rabiei E, White M, Mosleh A, Lyer S, Woo J (2018) Component reliability modeling through the use of Bayesian networks and applied physics-based models. In: 2018 annual reliability and maintainability symposium (RAMS), pp 1–7
Schwarz G (1978) Estimating the dimension of a model. Ann Stat
Scutari M (2009) Learning Bayesian networks with the bnlearn r package. arXiv:0908.3817
Spirtes P, Glymour CN, Scheines R, Heckerman D (2000) Causation, prediction, and search. MIT Press, Cambridge
Tomek I (1976) An experiment with the nearest-neighbor rule. IEEE Trans Syst Man Cybern SMC–6(6):448–452
Tsamardinos I, Aliferis CF, Statnikov A (2003) Time and sample efficient discovery of Markov blankets and direct causal relations. In: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining, pp 673–678
Tsamardinos I, Aliferis CF, Statnikov AR, Statnikov E (2003) Algorithms for large scale Markov blanket discovery. FLAIRS Conf 2:376–380
Tsamardinos I, Brown LE, Aliferis CF (2006) The max-min hill-climbing Bayesian network structure learning algorithm. Mach Learn 65(1):31–78
Verma T, Pearl J (1991) Equivalence and synthesis of causal models. UCLA, Computer Science Department
Weiss GM (2004) Mining with rarity: a unifying framework. SIGKDD Explor Newsl 6(1):7–19. https://doi.org/10.1145/1007730.1007734
Wilson DL (1972) Asymptotic properties of nearest neighbor rules using edited data. IEEE Trans Syst Man Cybern 2(3):408–421. https://doi.org/10.1109/TSMC.1972.4309137
Yen SJ, Lee YS (2009) Cluster-based under-sampling approaches for imbalanced data distributions. Expert Syst Appl 36(3, Part 1):5718–5727. https://doi.org/10.1016/j.eswa.2008.06.108
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Dharmasena, I., Domaratzki, M. & Muthukumarana, S. Modeling mobile apps user behavior using Bayesian networks. Int. j. inf. tecnol. 13, 1269–1277 (2021). https://doi.org/10.1007/s41870-021-00699-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s41870-021-00699-7