Skip to main content
Log in

Modeling mobile apps user behavior using Bayesian networks

  • Original Research
  • Published:
International Journal of Information Technology Aims and scope Submit manuscript

Abstract

Modern apps based businesses are increasingly interested in data driven decision making to achieve business goals as well as retaining their customer base. In this paper, we propose a Bayesian network approach to assess the mobile apps user behavior. We propose a strategy to build Bayesian networks and further improve the causal networks using re-sampling methods to best represent the causal representation between app user retention and in-app features. Structural hamming distances (SHD) are then used for assessing similar Bayesian network structures learned using data available from a local mobile app developing company. We also conduct a simulation study to assess the effect of re-sampling techniques towards the Bayesian network performance with various learning algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Acid S, de Campos LM (2003) Searching for Bayesian network structures in the space of restricted acyclic partially directed graphs. J Artif Int Res 18(1):445–490

    MathSciNet  MATH  Google Scholar 

  2. Akaike H (1974) A new look at the statistical model identification. IEEE Trans Autom Control 19(6):716–723

    Article  MathSciNet  Google Scholar 

  3. Ashcroft M (2012) Bayesian networks in business analytics. In: Proceedings of the federated conference on computer science and information systems, pp 955–961

  4. Barua S, Islam MM, Yao X, Murase K (2014) MWMOTE—majority weighted minority oversampling technique for imbalanced data set learning. IEEE Trans Knowl Data Eng 26(2):405–425. https://doi.org/10.1109/TKDE.2012.232

    Article  Google Scholar 

  5. Bobek S, Baran M, Kluza K, Nalepa GJ (2013) Application of Bayesian networks to recommendations in business process modeling. In: AIBP@ AI* IA, pp 41–50

  6. Bozdogan H (1987) Model selection and Akaike’s information criterion (aic): the general theory and its analytical extensions. Psychometrika 52(3):345–370

    Article  MathSciNet  Google Scholar 

  7. Brown LE, Tsamardinos I, Aliferis CF (2004) A novel algorithm for scalable and accurate Bayesian network learning. In: Medinfo, pp 711–715

  8. de Campos LM (2006) A scoring function for learning Bayesian networks based on mutual information and conditional independence tests. J Mach Learn Res 7:2149–2187

    MathSciNet  MATH  Google Scholar 

  9. Chakraborty S, Mengersen K, Fidge C, Ma L, Lassen D (2016) A Bayesian network-based customer satisfaction model: a tool for management decisions in railway transport. Decis Anal 3(1):4

    Article  Google Scholar 

  10. Chawla N, Japkowicz N, Kołcz A (2004) Editorial: special issue on learning from imbalanced data sets. SIGKDD Explor 6:1–6. https://doi.org/10.1145/1007730.1007733

    Article  Google Scholar 

  11. Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP (2002) SMOTE: synthetic minority over-sampling technique. Tech. rep

  12. Chickering D, Geiger D, Heckerman D (1995) Learning Bayesian networks: search methods and experimental results. In: Preliminary papers of the fifth international workshop on artificial intelligence and statistics

  13. Chickering DM (1996) Learning Bayesian Networks is NP-Complete, learning from data: artificial intelligence and statistics v edn. Springer-Verlag. https://www.microsoft.com/en-us/research/publication/learning-bayesian-networks-is-np-complete/

  14. de Jongh M, Druzdzel MJ (2009) A comparison of structural distance measures for causal bayesian network models. Recent advances in intelligent information systems, challenging problems of science, computer science series, pp 443–456

  15. Dumais S, Platt J, Heckerman D, Sahami M (1998) Inductive learning algorithms and representations for text categorization. In: Proceedings of the seventh international conference on information and knowledge management, CIKM ’98. Association for Computing Machinery, New York, pp 148–155. https://doi.org/10.1145/288627.288651

  16. Ezawa K, Singh M, Norton SW (1996) Learning goal oriented Bayesian networks for telecommunications risk management. In: Proceedings of the 13th international conference on machine learning. Morgan Kaufmann, pp 139–147

  17. Fawcett T, Provost F (1996) Combining data mining and machine learning for effective user profiling. In: Proceedings of the second international conference on knowledge discovery and data mining, KDD’96. AAAI Press, pp 8–13

  18. Fernández A, García S, Herrera F, Chawla NV (2018) SMOTE for learning from imbalanced data: progress and challenges, marking the 15-year anniversary. https://doi.org/10.1613/jair.1.11192

  19. Haibo H, Yang B, Garcia EA, Shutao L (2008) ADASYN: adaptive synthetic sampling approach for imbalanced learning. In: 2008 IEEE international joint conference on neural networks (IEEE world congress on computational intelligence). IEEE, pp 1322–1328. https://doi.org/10.1109/IJCNN.2008.4633969. http://ieeexplore.ieee.org/document/4633969/

  20. Han H, Wang WY, Mao BH (2005) Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. Tech. rep

  21. Hart P (1968) The condensed nearest neighbor rule (corresp.). IEEE Trans Inf Theory 14(3):515–516

    Article  Google Scholar 

  22. Heckerman D, Geiger D, Chickering D (1995) Learning Bayesian networks: the combination of knowledge and statistical data. Mach Learn 20:197–243. https://doi.org/10.1007/BF00994016

    Article  MATH  Google Scholar 

  23. Hibbert D, Armstrong N (2009) An introduction to Bayesian methods for analyzing chemistry data: part II: a review of applications of Bayesian methods in chemistry. Chemometr Intell Lab Syst 97(2):211–220

    Article  Google Scholar 

  24. Hoch D (2014) App retention improves—apps used only once declines to 20%. http://info.localytics.com/blog/app-retention-improves

  25. Kubat M, Holte RC, Matwin S (1998) Machine learning for the detection of oil spills in satellite radar images. Mach Learn 30(2):195–215. https://doi.org/10.1023/A:1007452223027

    Article  Google Scholar 

  26. Kubat M, Matwin S (1997) Addressing the curse of imbalanced training sets: one-sided selection. In: ICML

  27. Larranaga P, Poza M, Yurramendi Y, Murga RH, Kuijpers CMH (1996) Structure learning of Bayesian networks by genetic algorithms: a performance analysis of control parameters. IEEE Trans Pattern Anal Mach Intell 18(9):912–926

    Article  Google Scholar 

  28. Laurikkala J (2001) Improving identification of difficult small classes by balancing class distribution. In: Proceedings of the 8th conference on AI in medicine in Europe: artificial intelligence medicine, AIME ’01. Springer, Berlin, pp 63–66

  29. Lewis DD, Catlett J (1994) Heterogeneous uncertainty sampling for supervised learning. In: Cohen WW, Hirsh H (eds) Machine learning proceedings 1994. Morgan Kaufmann, San Francisco, pp 148–156. https://doi.org/10.1016/B978-1-55860-335-6.50026-X

  30. Lewis DD, Ringuette M (1994) A comparison of two learning algorithms for text categorization

  31. Lucas P (2001) Bayesian networks in medicine: a model-based approach to medical decision making

  32. Margaritis D (2003) Learning Bayesian network model structure from data. Carnegie-Mellon Univ Pittsburgh Pa School of Computer Science, Tech. rep

  33. Neapolitan R (2003) Learning Bayesian networks. Prentice Hall. https://doi.org/10.1145/1327942.1327961

  34. Neapolitan RE et al (2004) Learning Bayesian networks, vol 38. Pearson Prentice Hall, Upper Saddle River

    Google Scholar 

  35. Needham CJ, Bradford JR, Bulpitt AJ, Westhead DR (2007) A primer on learning in Bayesian networks for computational biology. PLoS Comput Biol 3(8):1–8. https://doi.org/10.1371/journal.pcbi.0030129

    Article  Google Scholar 

  36. Parsons S (2011) Probabilistic graphical models: principles and techniques. Knowl Eng Rev 26(2):237–238

    Article  Google Scholar 

  37. Pearl J (1988) Probabilistic reasoning in intelligent systems. In: Probabilistic reasoning in intelligent systems. Morgan Kaufmann. https://doi.org/10.1016/B978-0-08-051489-5.50001-1

  38. Perrier E, Imoto S, Miyano S (2008) Finding optimal Bayesian network given a super-structure. J Mach Learn Res 9(Oct):2251–2286

    MathSciNet  MATH  Google Scholar 

  39. Perro J (2018) Mobile apps: what’s a good retention rate? http://info.localytics.com/blog/mobile-apps-whats-a-good-retention-rate

  40. Provost F, Fawcett T (2001) Robust classification for imprecise environments. Mach Learn 42(3):203–231. https://doi.org/10.1023/A:1007601015854

    Article  MATH  Google Scholar 

  41. Rabiei E, White M, Mosleh A, Lyer S, Woo J (2018) Component reliability modeling through the use of Bayesian networks and applied physics-based models. In: 2018 annual reliability and maintainability symposium (RAMS), pp 1–7

  42. Schwarz G (1978) Estimating the dimension of a model. Ann Stat

  43. Scutari M (2009) Learning Bayesian networks with the bnlearn r package. arXiv:0908.3817

  44. Spirtes P, Glymour CN, Scheines R, Heckerman D (2000) Causation, prediction, and search. MIT Press, Cambridge

    MATH  Google Scholar 

  45. Tomek I (1976) An experiment with the nearest-neighbor rule. IEEE Trans Syst Man Cybern SMC–6(6):448–452

    MathSciNet  MATH  Google Scholar 

  46. Tsamardinos I, Aliferis CF, Statnikov A (2003) Time and sample efficient discovery of Markov blankets and direct causal relations. In: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining, pp 673–678

  47. Tsamardinos I, Aliferis CF, Statnikov AR, Statnikov E (2003) Algorithms for large scale Markov blanket discovery. FLAIRS Conf 2:376–380

    Google Scholar 

  48. Tsamardinos I, Brown LE, Aliferis CF (2006) The max-min hill-climbing Bayesian network structure learning algorithm. Mach Learn 65(1):31–78

    Article  Google Scholar 

  49. Verma T, Pearl J (1991) Equivalence and synthesis of causal models. UCLA, Computer Science Department

  50. Weiss GM (2004) Mining with rarity: a unifying framework. SIGKDD Explor Newsl 6(1):7–19. https://doi.org/10.1145/1007730.1007734

    Article  Google Scholar 

  51. Wilson DL (1972) Asymptotic properties of nearest neighbor rules using edited data. IEEE Trans Syst Man Cybern 2(3):408–421. https://doi.org/10.1109/TSMC.1972.4309137

    Article  MathSciNet  MATH  Google Scholar 

  52. Yen SJ, Lee YS (2009) Cluster-based under-sampling approaches for imbalanced data distributions. Expert Syst Appl 36(3, Part 1):5718–5727. https://doi.org/10.1016/j.eswa.2008.06.108

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Saman Muthukumarana.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dharmasena, I., Domaratzki, M. & Muthukumarana, S. Modeling mobile apps user behavior using Bayesian networks. Int. j. inf. tecnol. 13, 1269–1277 (2021). https://doi.org/10.1007/s41870-021-00699-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s41870-021-00699-7

Keywords

Navigation