Skip to content
BY 4.0 license Open Access Published by De Gruyter Open Access August 14, 2020

The predicted load balancing algorithm based on the dynamic exponential smoothing

  • Lijie Yan and Xudong Liu EMAIL logo
From the journal Open Physics

Abstract

To a large extent, the load balancing algorithm affects the clustering performance of the computer. This paper illustrated the common load balancing algorithms and elaborated on the advantages and drawbacks of such algorithms. In addition, this paper provides a kind of balancing algorithm generated on the basis of the load prediction. Due to the dynamic exponential smoothing model, such an algorithm helps obtain the corresponding smoothing coefficient with the server node load time series of current phrase and allows researchers to make prediction with the load value at the next moment of this node. Subsequently, the dispatcher makes the scheduling with the serve request of users according to the load predicted value. OPNET Internet simulated software is applied to the test, and we may conclude from the results that the application of such an algorithm acquires a higher load balancing efficiency and better load balancing effect.

1 Introduction

As the computer network has been continuously upgraded and people access the Internet more and more frequently, servers of all sorts have afforded more and more pressure of being visited. Due to the increasing load, part of the servers has encountered certain issues, including but not limited to overtime, existing serious downtime, and incapacity of corresponding with the request of users. Fortunately, the computer cluster has appeared to better address the aforementioned issues. The computer cluster means that numerous heterogeneous or homogeneous computers are able to collaborate with each other to accomplish tasks or provide services via one certain means. With the help of the dispatcher, the cluster sends the request to each computer, which will effectively show the utilization ratio of servers. The load balancing algorithm [1,2] makes up an essential element of the computer cluster, having a greater effect on clustering performance. The current relevant research of the load balancing algorithm mainly concentrates on two aspects, namely, the dynamic balancing algorithm and the static balancing algorithm.

At present, the use of static balancing is more prevalent. The weighted polling algorithm [3,4] is a kind of static balancing algorithm, referring to the method of successively allocating the request of users to the server nodes. Such an algorithm regards all server nodes as same. The weighted polling algorithm takes the performance difference of the server node into consideration based on the polling algorithm and gives the weighted coefficient to each node. In addition, the weighted coefficient is used to represent the handling ability, namely, the weighted polling algorithm. By comparison, the other static balancing algorithms are rather simple, for instance, aiming address and random allocation algorithm, neither of which are difficult to implement. However, as they donot take the current actual loading condition of the server into consideration, the generating balancing effect is oftentimes unideal. The deadline constraint scheduler was designed to satisfy the user constraints. The algorithm aims to maximize the number of jobs that can be run in a cluster while satisfying the time requirements of all jobs [5]. A novel load balancing strategy is proposed through the use of a genetic algorithm. The users of the algorithm attempt to balance the load of the cloud infrastructure while trying to minimize the completion time of a given task set [6].

Therefore, the current relevant studies on the load balancing algorithm focus on the dynamic balancing algorithms, which rather typically include the fastest response speed algorithm and the least connection algorithm. The allocated node by the fastest response speed algorithm has the shortest corresponding time, whereas the new request of users is allocated to the serve node through the least connection algorithm. This kind of server node has rather little connection numbers. In addition, a dynamic feedback balancing algorithm [7,8] was eventually put forward. Such an algorithm takes the load characteristics of the server node as the reference, makes the adjustment with every weighted factor, and reflects the real-time load under the support of weighted value. [9] Researchers studied how dynamic controllers are deployed by adjusting the deployment location of controllers so as to limit the communication latency and by adjusting the number of controllers so as to support dynamic loads. Reference [10] leads to the proposal of the improved dynamic load balancing strategy based on feedback, which could attain equivalent performance at the expense of a smaller cost. Compared with the static balancing algorithm, a dynamic balancing algorithm helps acquire a better balancing effect and can enhance the performance to a certain extent. However, this algorithm requires less time interval and sometimes consumes a larger amount of server resources while sending the load message to the dispatcher dynamically. We have compared the recent dynamic load balancing algorithms in Table 1.

Table 1

Comparison of dynamic load balancing algorithms

Year Type of application Advantages/improvements Limitations
2018 [11] Real workload To improve the utilization of memory and to enhance revenue VM instances cost per hour are much higher in auction-based mechanism than the predicted (reserve) instance
2004 [12] Workflow applications To improve the makespan time and cost Algorithm does not scale well with complex issues
2019 [13] Independent tasks To improve the makespan time, throughput, availability, and cost Novel compromise solution between time and cost (conflicting objectives) is not discussed
2019 [14] Workflow applications To cut down cost and time Trade-off solution between conflicting objectives is not discussed
2019 [15] Independent applications To improve various influential parameters and to provide nondominance set of solutions using Pareto theory Proposed algorithm is not using any machine learning approach to predict the demand of users

2 Dynamic exponential smoothing prediction model

2.1 The calculation of server node load

The load reflects the current equipped performance of the server, and each server node is oftentimes heterogeneous in the computer cluster. This will make the allocation task of the dispatcher more balanced, show the largest utilization ratio of all the resources clearly, and guarantee that the current performance of the characterization server node will not encounter mistakes. In general, some indexes have an effect on the server load, which include the utilization of disk I/O and memory usage. Assuming I C, I D, I M, and I N to represent, respectively, the server node CPU, disk bandwidth utilization, memory utilization, and Internet bandwidth utilization, then the calculation formula of the load is

(1) L = ( φ 1 I C + φ 2 I M + φ 3 I D + φ 4 I N ) × 100 %

where φ i represents the coefficient with the weight of all the indexes; furthermore, q i = 1 .

For instance, when the streaming media server in the RTP contract is selected as the object and the FTP server is selected as the object, the weighted value of disk I/O utilization should be promoted to a level. In addition, the weighted value of bandwidth utilization should be promoted to a certain level. In the practical application, researchers have taken the particular emphasis of server node as the proof to adjust φ , which aims to get a better effect.

2.2 Single exponential smoothing model

The exponential smoothing model, a kind of short time series prediction method, is both concise and practicable. The model covers all the fields such as electricity and business. In addition, the secondary exponential smoothing and single exponential smoothing models are exponential smoothing methods since the higher level exponential smoothing algorithm brings larger pressure to the load dispatcher on the calculation, which is not good for the normal expression of computer cluster performance. Therefore, a single exponential smoothing model has been selected. Assuming {L t } to represent the time series of the server node load, and the load predicted value at the moment of t + 1 needs to be obtained as for the t moment. For the single exponential smoothing prediction model [16,17,18,19,20,21,22], it is represented in the following formula:

(2) S t = a L t + ( 1 a ) S t 1 ,

(3) L t + 1 = S t .

where S t represents the smoothing value at the moment of t, namely, the predicted value at the moment of t + 1, α represents the smoothing coefficient, and L t and S t−1, respectively, represent the load value at t moment and the smoothing value of t − 1 moment of the server node; unfolding it with the recursive method, and we may obtain the following formula:

(4) S t = i = 1 t α ( 1 α ) t i L i + ( 1 α ) t S 0

In this formula, S 0 refers to the initial smoothing value. It can be obtained from Eq. (4) that the predicted value S t at t moment involves the use of all the historical data {L t }, in addition, aiming at the specific load time series {L t }, the smoothing initial value S 0, and the smoothing coefficient α affect the predicted value at t moment to a great extent.

S 0 and α represent themselves with the static model in the traditional exponential smoothing model. We have only discussed smoothing coefficient α while ignoring the influence of smoothing initial value S 0, and what is definite is that α affects the predicted result accuracy to a larger degree. The most important point of the exponential smoothing model is to find the reasonable smoothing coefficient α.

As for the static smoothing coefficient α, some means of improvement are given. In fact, it is not comprehensive if one attempts to find the optimal static smoothing coefficient, and certain conflicts will arise with the load time series {L t }. On the server node, the length of time series {L t } of load prediction is always becoming ever longer, and this changing trend is unpredictable, which leads to the attempting method to obtain the optimal smoothing coefficient α without any availability. Therefore, one kind of the prediction model of dynamic smoothing coefficient is put forward.

2.3 Dynamic exponential smoothing model

The exponential smoothing method [23,24] is easy to be operated and possesses rather stronger adaptability, and its compiling procedure is not complicated. However, it has increasingly imposed an effect on the practical application. α exponential smoothing method shows the historical data by means of historical information, which covers the entire model. What’s more, it follows the rule that the closer with the current moment means the larger information weighted. Such an idea of revising the historical data weight is relatively consistent with the actual situation. In general, the more recent information will have a higher value, and viewing from a certain extent, it reflects the practical rules. The exponential smoothing method [20,21,22] has a wider range of usage and leads to a gradually increasing utilization ratio. Therefore, the research of the exponential smoothing method becomes more significant. However, during the specific usage, there exists a rather prominent issue in the exponential smoothing method, namely, the selection of the smoothing coefficient α. In the practical application, it should be selected in relation to the experience. When the data have a smaller fluctuation, it has usually been selected between 0.1 and 0.3; if the data have a bigger fluctuation, then it has usually been selected between 0.6 and 0.8.

Since the selection standard of α is not very clear, some problems may appear in the practical application. Due to the arising problems, the accuracy of the model prediction will undoubtedly be affected, which is not good for the people lacking experience. Given the existed error, we may then construct the dynamic exponential smoothing model, and we may select the predicted value and practical value for the aiming function. We can construct a two-dimension planning model and use this model to obtain the dynamic answer of the exponential smoothing method, though this kind of parameter is subject to changes as time goes by. Through comparison by means of practical examples, we may conclude that this method has a higher accuracy, which is contributory to ensure the smoothing coefficient α. If L t+1 is the predicted value at t + 1 moment, then the predicted model of dynamic smoothing coefficient is

(5) L t + 1 = S t = i = 1 t α t ( 1 α t ) t i L t + ( 1 α t ) t S 0

where α t represents the optimal dynamic smoothing coefficient at t moment on the basis of the load time series {L t }, and at the same time of the continuing longer of {L t }, α t makes the change in order to correspond with the load time series {L t } and represents the dynamic characteristic. When making an assessment with the optimal dynamic smoothing coefficient on the basis of the load time series {L t }, the raised assessing model can predict the error square and regards the minimal SSE as the target, namely,

(6) MinSSE = i = 1 t ( L i L ¯ i ) 2 = i = 1 t ( L i S i 1 ) 2

then put the formula (4) into it to obtain:

(7) MinSSE = i = 1 i L i i = 1 i α t ( 1 α t ) t i L i ( 1 α t ) i S 0 2

Then, the optimal smoothing coefficient α t can be obtained through the formula (7), and while solving the similar nonlinear optimizing model, we should adapt the steepest descent method with the quicker disappearing rate to the model. At first, we should set the smoothing coefficient initial value α t 0 , ε > 0 , and take it as the control condition:

Step 1, aim at the given smoothing coefficient α t i (the value of i is 0, 1, 2, etc.), and then calculate SS E ( α t i ) . If | SS E ( a t i ) | ε , then α t i is the similar optimal answer. Step 2, assume | SS E ( α t i ) | > ε , and then calculate with the optimal step λ i (the value of i is 0, 1, 2, etc.) with the golden section method in the one-dimension exploration, in the meanwhile, assume α t i = a t i λ × SS E ( α t i ) , then go back to the first step. And SS E ( a t i ) = dSSE dat | α t α t i . Acquire the optimal dynamic smoothing coefficient based on the formula (7) and construct the dynamic smoothing prediction model of formula (5) on this basis, into which we can’t directly apply the load balancing algorithm.

At first, the load time series {L t } is becoming ever longer, so both of the times of formula (5, 7) are increasing, and the complexity of this algorithm is increasingly higher. In addition, the load dispatcher needs to retain lots of time series {L t }, which makes it more time-consuming on obtaining the optimal smoothing coefficient. Thus, the computer cluster is not corresponding with the given purpose. Next, α in the dynamic smoothing coefficient is not the same with the one in the traditional static smoothing coefficient, due to the existence of the limit of 0 < α < 1.

The existing load time series {L t } is sharply declining or increasing with the condition of α > 1; in fact, in the {L t } with concave or convex type, the condition of the smoothing coefficient α < 0 might appear. Under the above circumstances, the application of the exponential smoothing method cannot manifest the thinking of “laying more emphasis on the closer than the farther,” and it is not necessary to consider the smoothing initial value S 0 according to formula (5). In addition, to connect with the definition of load in formula (1), there surely exists 0 < Li < 100. Therefore, while the load time series {L t } is continuously increasing, there always exists the trend of sharply declining or increasing, or viewing from the entirety, the probability of concave or convex type is gradually becoming smaller. It can be assumed that the load time series {L t } achieved the certain length, and the smoothing coefficient α is up to 0 < α < 1. In this way, the time series {L tt0} with the sufficient distance from the current t moment cannot manifest its importance. Given the above condition, the dynamic smoothing prediction model through the improvement is put forward:

(8) L t + 1 = S t = i = t + 1 t 0 t α t ( 1 α t ) t i L t + ( 1 α t ) t S t t 0

In this formula, t 0 expresses the length of retained load time series {L t }. Furthermore, we can revise the assessment model of solving the optimal smoothing coefficient αt as

(9) MmiSSE = i = t + 1 t 0 t L i i = 1 i α t ( 1 α t ) t i L i ( 1 α t ) i S t t 0 2

Due to the improved model and the loss of the load time series prior to the moment of tt0, both of the times of formula (8) and (9) are increasing while load time series {L t } are continuously changing, which helps keep at the constant time square relevant with t0 and make the complexity of algorithm declining further. In addition, when load time series {L t } reaches a certain length, the accepted effect of smoothing initial value S 0 is gradually smaller, ending with the loss. Thus, the static value can be regarded as the smoothing initial value, and its definition based on this perception is

(10) S 0 = 1 t 0 i = 1 t 0 L i

The most critical point of the dynamic smoothing prediction model obtained through the improvement is to select the appropriate retained series length t 0 and to ensure that the application of algorithm reaches the prediction accuracy and that complexity strikes a rational balance.

Figure 1 shows the result of the dynamic exponential smoothing prediction model obtained through the improvement on the load prediction. The actual load value refers to the actual result obtained from the specific one server through the collection with an interval of one minute. Viewing t 0 = 5 and t 0 = 10 as the selected points, and when t 0 = 5, the prediction starts with the sixth load value; according to the similar analogy, when t 0 = 10, the prediction starts with the eleventh load value. Through the analysis of Figure 1, it is not hard to find that both of the predicted results of t 0 = 5 and t 0 = 10 meet with the actual load curve. Furthermore, when t 0 = 10, the improving predicted accuracy is not prominent, and comparing with t 0 = 5, this algorithm has a higher degree of complexity. Therefore, with respect to the dynamic load prediction model, it is more reasonable to select the load time series retained length t 0 = 5.

Figure 1 
                  The predicted result of the dynamic exponential smoothing method on the load.
Figure 1

The predicted result of the dynamic exponential smoothing method on the load.

It can be seen from Figure 2 that the prediction model of the static smoothing coefficient led to the load predicted result. Due to the varieties of the changing of load time series, it is extremely complicated for the single smoothing coefficient to adapt the entire time series, and thus, it cannot lead to a higher accuracy of prediction. Obviously, judging from the comparison with the above two figures, the predicted result of the dynamic exponential smoothing method is superior to the result of the static exponential smoothing method.

Figure 2 
                  The predicted result of the static exponential smoothing method on the load.
Figure 2

The predicted result of the static exponential smoothing method on the load.

3 Load balancing algorithm

3.1 The description of load predicted function

Assume the serve node as { S v i } and the server number of computer cluster is expressed with n. If the specific server starts, then it will collect the self-load information every one-minute interval, and we may generate the initial time series with the load prediction { L t 0 = 5 } S v i before allocating this kind of series to the dispatcher. The dispatcher stores such a time series and helps obtain the smoothing initial value S S v i of S v i through formula (10). Assume the obtaining function of predicted value as S s v i = Exp prediction( L s v i ) , then the specific working process of this function is as follows:

(1) As for the server node S v i , the dispatcher finds its load time series { L t 0 = 5 } S v i and discards the earliest load value linearly in this time series while adding the latest load value L S v i at the end of time series { L t 0 = 5 } S v i . (2) Based on the smoothing value S s v i of the former moment and the new load time series { L t 0 = 5 } S v i as well as the dynamic smoothing coefficient assessment model of formula (9), we can obtain the optimal smoothing coefficient αt through the steepest descent method, which is corresponding with the load series. (3) In the dynamic smoothing prediction model of formula (8), we can put S s v i , the load time series { L t 0 5 } S v i and dynamic smoothing coefficient αt into the formula, and obtain the new smoothing value S s v i . We should no longer use the former smoothing value S s v i , but S s v i is regarded as the predicted value at the next moment of the server node S v i .

3.2 The description of load balancing algorithm

The user sends service request and the dispatcher accepts the request, and according to the load balancing algorithm, the request is sent to the optimal server node Sv i of the load to complete the process. The algorithm procedure is set out as follows:

(1) The user sends the service request and the dispatcher accepts the request. Then, select the minimal server node Sv i for the minimal Sv i to express the minimal lad of the next moment node S v i . (2) The dispatcher sends the request to the server node Sv i and makes the settlement. (3) After the server node Sv i settled the request, it will bring the load data of itself current moment to the response, and in the meantime, it sends the response to the dispatcher. (4) The dispatcher receives the response from Sv i and obtains the load S v i at the new moment, scheduling the load prediction function to dispatch the load function Exprediction( L s v i ) and acquire the load predicted value S v i of next moment to instruct the next scheduling of dispatcher.

4 The internet stimulated result and analysis

In order to test the characteristic of algorithm, the Internet stimulated software OPNET15.0 is applied, and this software integrates the application and stimulation, supporting lots of Internet communication contract. In addition, the software is able to dispatch the imitated system, stimulate the separating matter, and analyze the systematic performance and behavior. OPNET provides the modeling mechanism of process layer, node layer, and internet layer. Among them, there exists the load balancing algorithm in the process layer of the dispatcher. The server group is in the computer cluster, and it consists of three web servers. According to the performance ratio of 4:7:10, each server connects the dispatcher by means of 100 M line. Furthermore, the client consists of six subnets, five internal subnets, and one external subnet. In addition, each subnet covers 45 subscriber terminals. First, we should construct the environment, and then select the HTTP application in the stimulated setting, imitate the client to send HTTP to the server and keep the stimulated time of 5 h. In order to test the stimulated characteristic of the algorithm, we should make a comparison between this algorithm and the least connection algorithm.

Judging from whether load balancing is based on the current state of the system, algorithms can be divided into two categories: static load balancing algorithms and dynamic load balancing algorithms. This classification method is commonly used in practice.

The static load balancing algorithm judges the system load state according to the current state of the system, and the destination of the heavily loaded virtual machine migration is based on the priority of the virtual machine set in advance, not the load state of the destination virtual machine. Such algorithms as round-robin scheduling and weighted round-robin scheduling fall into the category of static load balancing algorithms (Figure 3).

Figure 3 
               Comparison of time series with high correlation coefficient on the load.
Figure 3

Comparison of time series with high correlation coefficient on the load.

The dynamic load balancing algorithm uses the current state of the system to determine the load state of the system. The destination of the heavily loaded virtual machine migration is not based on the priority of the virtual machine set in advance, but also the load state of the virtual machine cluster where the virtual machine is located. The lightly loaded virtual machine is selected as the migration destination. The dynamic load balancing algorithm is both efficient and flexible.

For instance, for the utilization of CPU, we should make use of the least connection algorithm to compare with the utilization of CPU and the one obtained from the writer’s algorithm of each server. Its utilization of CPU stabilizes at 10%, 17%, 25%, and 20%. Our research findings show that by means of the least connection algorithm, the dispatcher didn’t respond in accordance with the load ability of the server node on the allocation task, leading to the higher CPU utilization ratio of part nodes and the unbalanced load. On the basis of the balancing algorithm of the writer, all of them uses the server node resource in balance, which makes it possible to ensure the load balance. In conclusion, compared with the least connection algorithm, the balancing algorithm of the writer obviously improved the balanced effect (Figures 4 and 5).

Figure 4 
               Comparison of time series with low correlation coefficient on the load.
Figure 4

Comparison of time series with low correlation coefficient on the load.

The running times of five distinct load prediction algorithms are compared in Figure 5. Among them, Load_pred is the algorithm proposed in this paper. It can be seen from the figure that the Load_pred algorithm has the shortest running time when it makes predictions at different moments. At the same time, the Load_pred algorithm has a very stable running time.

Figure 5 
               Running time.
Figure 5

Running time.

As can be seen from Figure 6, Load_pred has the lowest mean absolute percentage error (MAPE), indicating that it has good predictive stability.

Figure 6 
               Various algorithms for MAPE.
Figure 6

Various algorithms for MAPE.

5 Conclusion

The load predicted balancing algorithm acquired through the improvement allows one to construct the dynamic smoothing exponent and to obtain the load prediction with a higher accuracy through the comparison of short load time series. In addition, the algorithm allows the dispatcher to provide more accurate service demand for users and enhance the resource utilization ratio of the server node to a higher level. Using the OPNET Internet stimulated software, this algorithm connects the algorithm least, and it is obvious from the feedback result that the load balancing effect of this algorithm is superior. In the end, there inevitably exist some limitations in this study, and the insufficiency still needs to be discussed further. Therefore, subsequent related studies on the load balancing algorithm based on the dynamic exponential smoothing prediction will still be the hot topic.

Based on the load balancing scheduling algorithm, we have also taken into account such strategies as the differentiated resource collocation strategy and the feedback able load balancing strategy, and based on task classification and forecasting task resource demand, we have taken into consideration the system load balancing factors in the task scheduling process. However, the accurate type of task classification and task resource demand prediction holds key to this algorithm. The above two aspects shall be the starting points for future studies on the optimization of task scheduling algorithms based on load balancing.

Acknowledgments

The authors gratefully acknowledge three anonymous referees whose comments significantly contributed to the final version of this paper.

References

[1] Zhang YF, Wei QL, Zhao Y. The load balancing algorithm based on the load weighted value. Appl Study Comput. 2012;12:4711–3.Search in Google Scholar

[2] Wei QL. The research of dynamic feedback load balancing algorithm based on the cluster. Chongqing, China: Chongqing University; 2013.Search in Google Scholar

[3] Meng LM, Pan JX. The design of load balancing algorithm in the video monitoring system. J Zhejia Univ Technol. 2014;6:607–11.Search in Google Scholar

[4] Zhou BF. The prediction and application of dynamic exponential smoothing model. Nat Sci J Haerbin Norm Univ. 2013;4:25–27.Search in Google Scholar

[5] Kc K, Anyanwu K. Scheduling Hadoop jobs to meet deadlines. 2010 IEEE Second International Conference on Cloud Computing Technology and Science, Indianapolis, USA.10.1109/CloudCom.2010.97Search in Google Scholar

[6] Dasgupta K, Mandal B, Dutta P, Mandal JK, Dame S. A Genetic Algorithm (GA) based load balancing strategy for cloud computing. Procedia Technol. 2013;10:340–7.10.1016/j.protcy.2013.12.369Search in Google Scholar

[7] Meng LM, Xu Y. The load balancing algorithm based on the dynamic exponential smoothing prediction. J Zhejia Univ Technol. 2016;4:379–82.Search in Google Scholar

[8] Zhang Y, Liu G, Luan W, et al. An approach to class imbalance problem based on stacking and inverse random under sampling methods. Networking, Sensing and Control (ICNSC), 2018 IEEE 15th International Conference on. IEEE, 2018, 1–6.10.1109/ICNSC.2018.8361344Search in Google Scholar

[9] Huque MTIU, Si WS, Jourjon G, Gramoli V. Large-scale dynamic controller placement. IEEE Trans Netw Serv Manage. 2017;14(1):63–76.10.1109/TNSM.2017.2651107Search in Google Scholar

[10] Chen C, Diao TL, Wang WF. Improved dynamic load balancing strategy based on feedback. Comput Eng. 2010;36(14):34–36.Search in Google Scholar

[11] Zhang Y, Liu G, Luan W, et al. Application of SIRUS in Credit Card Fraud Detection. International Conference on Computational Social Networks. Cham: Springer; 2018. p. 66–78.10.1007/978-3-030-04648-4_6Search in Google Scholar

[12] Phua C, Alahakoon D, Lee V. Minority report in fraud detection: classification of skewed data. Acm Sigkdd Explorations Newsl. 2004;6(1):50–9.10.1145/1007730.1007738Search in Google Scholar

[13] Xiong Z, Wu Y, Ye C, Zhang X, Xu F. Color image chaos encryption algorithm combining CRC and nine palace map. Multimed Tools Appl. 2019;78(22):31035–55.10.1007/s11042-018-7081-3Search in Google Scholar

[14] Baskonus HM, Bulut H, Sulaiman TA. New complex hyperbolic structures to the Lonngren-wave equation by using Sine-Gordon expansion method. Appl Math Nonlinear Sci. 2019;4(1):129–38.10.2478/AMNS.2019.1.00013Search in Google Scholar

[15] Voit M, Meyer-Ortmanns H. Predicting the separation of time scales in a heteroclinic network. Appl Math Nonlinear Sci. 2019;4(1):279–88.10.2478/AMNS.2019.1.00024Search in Google Scholar

[16] Polikar R. Ensemble Learning. Scholarpedia. 2009;4(1):1–34.10.1007/978-1-4419-9326-7_1Search in Google Scholar

[17] Steve AlbrechtW. Albrecht, Chad, Albrecht, Conan C. Current Trends in Fraud and its Detection. Inf Syst Security. 2008;17(1):2–12.10.1080/19393550801934331Search in Google Scholar

[18] Hawkins DM. The problem of overfitting. J Chem Inf computer Sci. 2004;44(1):1–12.10.1021/ci0342472Search in Google Scholar PubMed

[19] Liu YQ, Wang C, Zhang L. Decision tree based predictive models for breast cancer survivability on imbalanced data. International Conference on Bioinformatics and Biomedical Engineering. Beijng, China: IEEE; 2009. p. 1–4.10.1109/ICBBE.2009.5162571Search in Google Scholar

[20] Zheng ZH, Wu XY, Srihari R. Feature selection for text categorization on imbalanced data. ACM SIGKDD Explorat Newsl. 2004;6(1):80–9.10.1145/1007730.1007741Search in Google Scholar

[21] Shatnawi R. Improving software fault-prediction for imbalanced data. International Conference on Innovations in Information Technology. Abu Dhabi, UAE: IEEE; 2012. 54–59.10.1109/INNOVATIONS.2012.6207774Search in Google Scholar

[22] He H, Garcia EA. Learning from Imbalanced Data. IEEE Trans Knowl Data Eng. 2009;21(9):1263–84.10.1109/TKDE.2008.239Search in Google Scholar

[23] Huang ZH, Xu X, Zhu HH, Zhou MC. An efficient group recommendation model with multiattention-based neural networks. IEEE Trans Neural Netw Learn Syst. 2020;1–14.10.1109/TNNLS.2019.2955567Search in Google Scholar PubMed

[24] Huang ZH, Xu X, Ni J, Zhu HH, Wang C. Multimodal representation learning for recommendation in Internet of Things. IEEE Internet Things J. 2019;6(6):10675–85.10.1109/JIOT.2019.2940709Search in Google Scholar

Received: 2020-03-30
Revised: 2020-06-04
Accepted: 2020-06-04
Published Online: 2020-08-14

© 2020 Lijie Yan and Xudong Liu, published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 27.4.2024 from https://www.degruyter.com/document/doi/10.1515/phys-2020-0108/html
Scroll to top button