Next Article in Journal
Flexible 5G New Radio LDPC Encoder Optimized for High Hardware Usage Efficiency
Previous Article in Journal
Flickering-Free Distance-Independent Modulation Scheme for OCC
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Memory-Efficient Deep Learning for Botnet Attack Detection in IoT Networks

1
Department of Engineering, Manchester Metropolitan University, Manchester M1 5GD, UK
2
Department of Computing and Mathematics, Manchester Metropolitan University, Manchester M1 5GD, UK
3
Department of Electrical and Information Engineering, Covenant University, Ota P.M.B. 1023, Nigeria
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(9), 1104; https://doi.org/10.3390/electronics10091104
Submission received: 1 April 2021 / Revised: 29 April 2021 / Accepted: 1 May 2021 / Published: 8 May 2021
(This article belongs to the Section Networks)

Abstract

:
Cyber attackers exploit a network of compromised computing devices, known as a botnet, to attack Internet-of-Things (IoT) networks. Recent research works have recommended the use of Deep Recurrent Neural Network (DRNN) for botnet attack detection in IoT networks. However, for high feature dimensionality in the training data, high network bandwidth and a large memory space will be needed to transmit and store the data, respectively in IoT back-end server or cloud platform for Deep Learning (DL). Furthermore, given highly imbalanced network traffic data, the DRNN model produces low classification performance in minority classes. In this paper, we exploit the joint advantages of Long Short-Term Memory Autoencoder (LAE), Synthetic Minority Oversampling Technique (SMOTE), and DRNN to develop a memory-efficient DL method, named LS-DRNN. The effectiveness of this method is evaluated with the Bot-IoT dataset. Results show that the LAE method reduced the dimensionality of network traffic features in the training set from 37 to 10, and this consequently reduced the memory space required for data storage by 86.49 % . SMOTE method helped the LS-DRNN model to achieve high classification performance in minority classes, and the overall detection rate increased by 10.94 % . Furthermore, the LS-DRNN model outperformed state-of-the-art models.

1. Introduction

The Internet of Things (IoT) and the Industrial IoT (IIoT) are part of the main information and communication technologies of the fourth industrial revolution (Industry 4.0) [1,2,3]. To a large extent, the global coronavirus (COVID-19) pandemic that started in 2020 has restricted human contacts and economic activities in most countries. In view of this, IoT and IIoT technologies will help in connecting people, homes, and businesses to a large-scale of computers, smart devices, sensors, vehicles, and industrial machines in smart cities. Unfortunately, IoT has become the primary target of malicious botnet (A botnet is a network of computing devices that are coordinated by a botmaster to perform specific tasks [4].) operators due to their proliferation and distributed nature. A large number of connected IoT devices are insecure because their default usernames and passwords remain unchanged [5,6]. A malicious botnet poses a serious cybersecurity threat to the Internet of Things (IoT) networks and their applications [7,8,9,10]. IoT devices are vulnerable to various botnet attacks such as Denial of Service (DoS), Distributed DoS (DDoS), Operating System (OS) fingerprinting, service scanning, data exfiltration, and keylogging [11]. Recently, a new IoT Peer-to-Peer (P2P) botnet, named HEH, exploited insecure Telnet services on ports 23 and 232 to wipe out all the data in IoT devices using the brute force method [12]. Therefore, IoT networks must be properly monitored and protected to detect and prevent cyberattacks.
Cybersecurity mechanisms such as encryption, authentication, and access control may not be strong enough to protect IoT networks against botnet attacks [13,14]. Therefore, an efficient Network Intrusion Detection System (NIDS) is needed to complement existing security mechanisms. NIDS will scan and monitor all the network traffic traces generated in IoT networks to detect botnet attacks. Signature-based NIDS can detect known attacks with high accuracy but it cannot identify zero-day (unknown) attacks. On the other hand, anomaly-based NIDS can detect unknown attacks but it has a higher False Positive Rate (FPR) than signature-based NIDS. Machine Learning (ML) method can be used to detect both known and unknown malicious network traffic traces in IoT networks [15,16,17,18,19]. Popular ML methods include Random Forest (RF), Support Vector Machine (SVM), Decision Tree (DT), k-Nearest Neighbour (kNN), Random Tree (RT), and Naive Bayes (NB). However, IoT networks generate massive network traffic data at very high velocity, and shallow neural networks cannot handle such big data because they have a limited number of trainable parameters.
Deep Learning (DL) is an advanced ML method that has more than one hidden layer in its neural network and it learns the feature representation of training data using multiple levels of abstraction [20,21,22]. Common DL architectures include Deep Neural Network (DNN), Convolutional Neural Network(CNN), Recurrent Neural Network (RNN), Deep Belief Network (DBN), Autoencoder (AE) and Restricted Boltzmann Machine (RBM). RNN is a type of neural network in which the connections between the neurons form a directed graph along a temporal sequence. Unlike other DL methods, it considers the temporal dependencies among the features in the training set. Most times, RNN is combined with fully-connected dense layer(s) to improve model’s classification performance. The integrated model architecture is referred to as Deep RNN (DRNN). In fact, DRNN models have been successfully applied to botnet detection in recent literature [11,21,22,23]. Therefore, we select DRNN model architecture because of its ability to model the temporal relationships among the features in network traffic data. However, for high feature dimensionality in the training data, a high network bandwidth and a large memory space will be needed to transmit and store the data, respectively in IoT back-end server or cloud platform for DL. Furthermore, given a highly imbalanced network traffic data, DRNN models will produce poor classification performance in minority classes.
In this paper, we propose a memory-efficient DL method, named LS-DRNN, for botnet attack detection in IoT networks. S-DRNN method employs SMOTE and DRNN algorithms only but LS-DRNN combines Long Short-Term Memory Autoencoder (LAE), SMOTE, and DRNN algorithms to achieve an efficient performance. AE is an unsupervised DL that is particularly suitable for feature representation as well as feature dimensionality reduction. There are different variants of AE including Stacked AE (SAE), Variational AE (AE), Sparse AE (SpAE), Convolutional AE (CAE), Deep AE (DAE), Denoising AE (DeAE), and Long Short-Term Memory AE (LAE). Unlike other variants of AE and similar to RNN, LAE uses Long Short-Term Memory (LSTM) to account for long-term dependencies among features while learning their representation and reducing the dimensionality. So, LAE is a good fit for feature dimensionality reduction in the botnet detection task. The main contributions of this paper are as follows:
  • LAE reduces the feature dimensionality of large-scale network traffic data using unsupervised DL method;
  • SMOTE generates additional samples for minority classes in low-dimensional network traffic data to achieve class balance;
  • DRNN performs multi-class classification of network traffic samples in balanced, low-dimensional data using supervised DL method;
  • DRNN, S-DRNN, and LS-DRNN models are trained, validated, and tested with network traffic samples in the Bot-IoT dataset [11], and their classification performance in 11-class classification scenario is evaluated.

2. Review of Related Works

In this section, we review related works to establish the novelty and the main contributions of this paper. Table 1 presents a summary of the state-of-the-art feature dimensionality reduction methods and class balance methods proposed for botnet detection in IoT networks.
Koroniotis et al. [11] used Pearson Correlation Coefficient (PCC) and joint entropy techniques to select the 10 most relevant features. Support Vector Machine (SVM), DRNN, and Long Short-Term Memory (LSTM) models were trained with these features to perform binary classification. The reduction in the number of features shortened the time taken to train the ML and DL models, but the classification performance was lower than when the full features were used for model training. Furthermore, the authors did not evaluate the performance of the feature selection method in a multi-class classification scenario. The same set of features was also used for ML-based intrusion detection in [24,25,26,27,28,29,30].
Kumar et al. [31] proposed a hybrid feature selection method, which combined PCC with Random Forest Mean Decrease Accuracy (RFMDA) and Gain Ratio (GR), to select the 10 most important features. Random Forest (RF), k-Nearest Neighbour (kNN), and Extreme Gradient Boosting (XGBoost) models were trained with these features to perform 5-class classification. Kumar et al. [32] used a mutual information-based feature selection method to select the 10 most relevant features. RF and XGBoost models were trained with these features for the 5-class classification task. Shafiq et al. [33,35] proposed a new feature selection algorithm based on the wrapper technique and Area Under Curve (AUC) metric. Then, C4.5, Naive Bayes (NB), RF, and SVM models were trained for 8-class classification. Koroniotis et al. [34] developed Multilayer Perceptron (MLP) and RNN models using 13 network traffic features. Asadi et al [36] proposed Particle Swarm Optimisation (PSO) algorithm to select 10 outstanding features. These features were used to train ML/DL models for binary classification. Popoola et al. [10] performed feature dimensionality reduction based on the LAE algorithm, while the Bidirectional LSTM (BLSTM) algorithm was used for the 5-class classification task. Other feature dimensionality reduction methods include Principal Component Analysis (PCA) in [37,38,39], and t-distributed Stochastic Neighbour Embedding (t-SNE) in [40].
Khan and Kim [48] proposed a hybrid intelligent model using both anomaly-based and misuse-based NIDS approaches. At the first stage, Logistic Regression (LR) and XGBoost algorithms were used to develop anomaly-based NIDS, while the LAE algorithm was used for misuse-based NIDS in the second stage of the system. The effectiveness of the hybrid model was evaluated with the ISCX-2012 data set. Roopak et al. [49] investigated the effectiveness of Multi-Layer Perceptron (MLP), CNN, LSTM, and CNN-LSTM models for DDoS attack detection in IoT networks. The authors simulated these models with the CICIDS2017 data set. Liaqat et al. [41] used the up-sampling method to increase the number of benign samples in the training data set. In [42,43,44,45], Synthetic Minority Oversampling Technique (SMOTE) method was used to generate additional samples for the minority classes. Mulyanto et al. [46] performed feature selection to reduce dimensionality while focal loss function was used to address class imbalance problem. Similarly, Injadat et al. [47] selected the most relevant features and additional minority samples were generated using SMOTE.

3. LS-DRNN: A Memory-Efficient Deep Learning Method

In this section, we explain how the LS-DRNN method is developed for botnet attack detection in IoT networks. LS-DRNN employs an unsupervised DL method (LAE), a sampling method (SMOTE), and a supervised DL method (DRNN).

3.1. LSTM Autoencoder

LAE method is an unsupervised DL method and it was used to reduce the dimensionality of network traffic features. Consequently, this process is expected to reduce the amount of memory space that will be required to store training data for DL in IoT back-end server or cloud platform.
A high-dimensional network traffic feature set is represented with X R n × a , where n is the total number of network traffic samples, and a is the feature dimensionality. This matrix was reshaped to form a sequential 3D tensor, X R n × 1 × a . LSTM is a recurrent neural network which learns latent space representation of network traffic features using input gate, forget gate, memory cell state, output gate, and hidden state. A single LSTM layer was used to reduce the dimensionality of the feature set based on Equations (1)–(6):
i 1 = σ r W i x 1 x + W i h 1 h 0 + b i 1 ,
f 1 = σ r W f x 1 x + W f h 1 h 0 + b f 1 ,
c 1 ˜ = σ h W c ˜ x 1 x + W c ˜ h 1 h 0 + b c ˜ 1 ,
c 1 = i 1 c 1 ˜ + f 1 c 0 ,
o 1 = σ r W o x 1 x + W o h 1 h 0 + b o 1 ,
h e n c = o 1 σ h c 1 ,
where i is the input gate vector; f is the forget gate vector; c is the memory cell state vector; o is the output gate vector; h is the hidden state vector; W ( · ) are the weight matrices; b ( · ) are the bias vectors; σ r is a sigmoid activation function; and σ h is a Rectified Linear Unit (ReLU). The weight matrices were initialised using He uniform technique [50].
Similarly, a single LSTM layer was also used to reconstruct the original high-dimensional features from the encoded, low-dimensional features based on Equations (7)–(12):
i 2 = σ r W i x 2 h e n c + W i h 2 h 0 + b i 2 ,
f 2 = σ r W f x 2 h e n c + W f h 2 h 0 + b f 2 ,
c 2 ˜ = σ h W c ˜ x 2 h e n c + W c ˜ h 2 h 0 + b c ˜ 2 ,
c 2 = i 2 c 2 ˜ + f 2 c 0 ,
o 2 = σ r W o x 2 h e n c + W o h 2 h 0 + b o 2 ,
h d e c = o 2 σ h c 2 .

3.2. Synthetic Minority Oversampling Technique

High class imbalance adversely affects the classification performance of ML/DL models. The class imbalance problem is usually handled by either under-sampling or over-sampling the data in the training set to achieve class balance. However, the under-sampling approach is not suitable for cases where the number of samples in one of the minority classes is very small (<10). In our case, there are only four samples in one of the minority classes (more details will be provided later in Section 4). Therefore, we simply ruled out the option of the under-sampling approach.
Recent studies recommended SMOTE as an efficient over-sampling method [42,43,44,45,47,51]. Therefore, SMOTE algorithm was proposed to deal with the high class imbalance problem in the training set in an 11-class classification scenario. Unlike the method in [52,53], which over-samples minority classes with replacement, the method employed in this paper generates synthetic examples by using techniques such as rotation and skew in order to achieve class balance [54].
These synthetic network traffic data were generated along the line segments joining any or all of the k nearest neighbours of the minority classes, where k  = 3 . Therefore, neighbours from the three nearest neighbours were randomly selected. The step-wise process of SMOTE is presented in Algorithm 1. The generation of synthetic samples (S) in the minority classes depends on the number of minority class samples (T), the over-sampling rate ( N % ), and the number of nearest neighbours (k). If N is less than 100 % , the minority class samples are randomised. We compute k nearest neighbours for each of the minority class only. This is a function of N, the current minority class sample (i), the integral multiples of 100 in N (j), and an array of random numbers ( n n _ a r r a y ). Z is an array of original minority class samples; r is the count of a number of synthetic samples generated, and V is an array of synthetic samples.
Algorithm 1: SMOTE Algorithm
Electronics 10 01104 i001

3.3. Deep Recurrent Neural Network

Given a low-dimensional network traffic feature set, X , and a corresponding ground truth label vector, y , the goal of DRNN is to learn the function that determines the output target whenever the input data sequences are presented. First, a new hidden state vector, h , is produced at each time step implementing an activation function on the previous hidden state, h 0 , and the current input, x .
Unlike Feedforward Neural Network (FNN), RNN has a hidden state which helps to model temporal dynamics of input data. RNN learns the temporal dynamics of a mini-batch of highly imbalanced network traffic features, X k , by transforming the input data and initial hidden state, h i n i t , with trainable parameters as stated in Equation (13):
h 1 k = σ h W x X k + W h h i n i t + b h ,
where h 1 k is the new hidden state when RNN is trained with the k t h mini-batch; W x and W h are the weights used for linear transformation of X k and h i n i t respectively; and b h is the bias. RNN layer output is further processed based on Equations (14)–(20) to produce DRNN layer output. Complete information about DRNN is presented in Algorithm 2.
Algorithm 2: DRNN Algorithm
Electronics 10 01104 i002
The hidden states of the four dense hidden layers are obtained by Equation (14):
h m k = σ h W h m h ( m 1 ) k + b m h
where m = [ 2 , 3 , 4 , 5 ] ; h m k is the hidden state of m t h hidden layer; h 1 k = h k ; W h m is the weight used for linear transformation of previous hidden state, h ( m 1 ) k ; b m h is the bias of m t h hidden layer; and σ h is a Rectified Linear Unit (ReLU) activation function given by Equation (15):
σ h ( a ) = m a x ( 0 , a ) .
If a is a negative value, the function returns 0; but the same a is returned when it is a positive value.
The hidden state of the fourth dense layer, h 5 k , is transformed by the dense output layer as stated in Equation (16):
y ˜ k = σ y W y h 5 n + b y ,
where n is the sample size of mini-batch of X and n = p / μ ; μ is the batch size and μ = 512 ; W y is the weight used for linear transformation of h 5 k ; b y is the bias of dense output layer; and σ y is the activation function of dense output layer.
In multi-class classification scenario, σ y is a softmax function given by Equation (17):
y ˜ k = e W y h 5 k + b y ω = 1 γ e W y h 5 k + b y ,
where γ is the number of classes in y . while the difference between y ˜ and y is measured by categorical cross-entropy loss function ( θ c ) in Equation (18):
L = θ c ( y k , y ˜ k ) = 1 n τ = 1 n ω = 1 γ y τ , ω log ( y ˜ τ , ω ) .
The performance of DRNN was validated with a different previously unknown highly imbalanced network traffic data, X v a , and its corresponding ground-truth labels, y v a . Training loss and validation loss are minimized in mini-batches over u epochs using an efficient first-order stochastic gradient descent algorithm named Adam in [55]. Trainable parameters of densely-connected DL model are represented by Equation (19):
Φ = W ( · ) , h ( · ) , b ( · ) .
For each epoch, Adam optimizer, ψ , updates Φ to minimize L as stated in Equation (20):
Φ = ψ Φ , L , α , β 1 , β 2 ,
where Φ is the new set of trainable parameters; α is the learning rate ( 0.0001 ); and β 1 and β 2 are the exponential decay rates ( 0.9 and 0.999 respectively).

4. Simulation and Performance Evaluation

In this section, we implement and evaluate the effectiveness of DRNN, S-DRNN, and LS-DRNN models with the Bot-IoT data set [11]. The research methodology involves data pre-processing, model development, and model evaluation, as shown in Figure 1.
Bot-IoT data set [11] contains the information about the features of 477 benign network traffic samples that were generated by five IoT scenarios, namely: a weather station, a smart fridge, motion-activated lights, a remote-controlled garage door, and a smart thermostat. Furthermore, the data set contains the features and the corresponding labels of 3,668,045 malicious network traffic samples that represent different botnet attack scenarios including DDoS, DoS, reconnaissance, and data theft. All the network traffic samples in the Bot-IoT dataset were grouped into 10 attack classes and a benign class namely: DDoS-HTTP (DDH), DDoS-TCP (DDT), DDoS-UDP (DDU), DoS-HTTP (DH), DoS-TCP (DT), DoS-UDP (DU), normal, OS Fingerprinting (OSF), Service Scanning (SS), Data Exfiltration (DE), and Key Logging (KL).
Data pre-processing stage involved input feature selection, label encoding, input feature normalization and random data splitting as described in [10]. The sample distribution of the BoT-IoT data set for training, validation and testing in Table 2 shows that the network traffic data is highly imbalanced across the 11 classes. All the 37 network traffic features have continuous values; and these values were scaled to a range of [ 0 , 1 ] using min-max transformation given by Equation (21):
X ˜ = X X m i n X m a x X m i n
where X m i n and X m a x are the minimum and maximum values of X respectively. Data pre-processing operations were implemented using Numpy, Pandas and Scikit-learn libraries developed for Python programming language.
Figure 2, Figure 3 and Figure 4 show the overview of the implementation of LAE, SMOTE, and DRNN methods, respectively. The classification performance of ML/DL model depends on the quality of the training data and the choice of the right network topology. Therefore, we performed extensive experimentation with different recurrent layer, dense layer, activation function, batch size, and epochs to determine the most suitable DRNN architecture. The optimal DRNN structure shown in Figure 4 has a single recurrent layer, four dense layers, and an output layer. There were 100 neurons each in the recurrent layer and the four dense layers. On the other hand, the number of neurons in the output layer depends on the number of classes in the training set. Hence, the number of neurons in the output layer was set to 11. ReLU activation function was used in the recurrent and dense layers, while softmax activation function was employed in the output layer. DRNN, S-DRNN, and LS-DRNN models were trained using mini-batch stochastic gradient descent algorithm [55], and we used a batch size of 64. These models were trained and validated for a period of 20 epochs using categorical cross-entropy loss function. Data processing were performed using open-source frameworks and libraries such as Pandas (https://pandas.pydata.org/), Numpy (https://numpy.org/), Sklearn (https://scikit-learn.org/stable/), and Imblearn (https://pypi.org/project/imblearn/). Model training, validation and testing were implemented using TensorFlow (https://www.tensorflow.org/) and Keras (https://keras.io/) frameworks developed for Python programming running on Ubuntu 16.04 LTS workstation with the following specifications: RAM (32 GB), processor (Intel Core i7-9700K CPU @ 3.60 GHz × 8), Graphics (GeForce RTX 2080 Ti/PXCIe/SSE2) and OS type (64-bit). The data sizes of the features in the training sets are analysed to determine the amount of storage space required for DL in memory-constrained IoT devices. Then, training losses and validation losses are analysed to evaluate the robustness of DL models against under-fitting and over-fitting, respectively. Lastly, the accuracy, precision, recall, F1 score, FPR, Negative Predictive Value (NPV), Balanced accuracy (BACC), Geometric Mean (GM), and Matthews Correlation Coefficient (MCC) of DL models are analysed to evaluate their classification performance. The time required to train DL models with the network traffic samples in the training sets and the time required to test the models with the network traffic samples in the testing sets is analysed to evaluate their training speed and detection speed, respectively.

5. Results and Discussion

In this section, we evaluate and compare the effectiveness of learning the discriminative features of benign network traffic and botnet attack traffic from highly imbalanced data using DRNN, S-DRNN, and LS-DRNN methods.

5.1. Results of DRNN Model

The data size of the feature set used for the training of the DRNN model was 1085.88 MB, as shown in Table 3. This implies that a large memory space (>1 GB) will be required to store training data for DL-based botnet attack detection in IoT back-end server or cloud platform. Figure 5 shows that the training loss of the DRNN model decreased from 0.1762 to 0.0027 as the number of epochs increased from 1 to 20. Similarly, its validation loss decreased from 0.0601 to 0.0036 , as shown in Figure 6. The consistent decrease in training loss and validation loss, as the number of epochs increased, implies that the DRNN model neither under-fit nor over-fit the network traffic data in the training and validation sets, respectively.
Table 4 shows that DRNN model achieved high accuracy (>99%), high NPV (>99%), and low FPR (≈0%) in each of the 11 classes. In addition, the precision, recall, F1 score, BACC, GM, and MCC of the DRNN model were high for the majority classes; however, the values of these performance metrics were relatively low for the minority classes. Consequently, the mean values of precision ( 88.58 % ), recall ( 88.81 % ), F1 score ( 88.63 % ), BACC ( 94.40 % ), GM ( 89.71 % ), and MCC ( 88.66 % ) were not very close to 100 % as expected. This means that high class imbalance in the training set degraded the ability of the DRNN model to correctly classify the network traffic samples in the minority classes. Furthermore, the time taken to train the DRNN model with 2,201,112 network traffic samples in the training set and the time taken to test the model with 1,467,410 samples in the testing set were 207.72 s and 0.7552 s, respectively.

5.2. Results of S-DRNN Model

SMOTE method was introduced to mitigate the adverse effect of high class imbalance on the classification performance of the DRNN model as earlier discussed in Section 5.1. This method generated a total of 4,612,442 synthetic network traffic samples in 548.72 s to achieve class balance across the 11 classes. Table 5 shows the distribution of network traffic samples that were used to train the S-DRNN model. Figure 5 shows that the training loss of the S-DRNN model decreased from 0.0917 to 0.0005 as the number of epochs increased from 1 to 20. Similarly, its validation loss decreased from 0.0473 to 0.0009 . Compared to the DRNN model, the initial training loss of the S-DRNN model reduced by 48 % , and its final training loss reduced by 83 % . Likewise, the initial validation loss of the S-DRNN model reduced by 21.19 % , and its final validation loss reduced by 76.08 % . Therefore, the S-DRNN model was more robust against under-fitting and over-fitting than the DRNN model. Table 4 and Table 6 shows that the classification performance of the S-DRNN model was better than that of the DRNN model, most especially in minority classes. Specifically, the SMOTE method that was introduced in the training of the S-DRNN model increased the:
  • precision by 4.96 % , recall by 14.65 % , F1 score by 10.07 % , BACC by 7.33 % , GM by 2.52 % , and MCC by 9.94 % , for DDH class;
  • precision by 7.43 % , recall by 5.23 % , F1 score by 6.34 % , BACC by 7.33 % , GM by 2.52 % , and MCC by 9.94 % , for DH class;
  • recall by 2.35 % , F1 score by 1.19 % , BACC by 1.18 % , and MCC by 1.18 % , for Norm class;
  • precision by 100 % , recall by 100 % , F1 score by 100 % , BACC by 50 % , GM by 100 % , and MCC by 100 % , for DE class; and
  • precision by 12.50 % , F1 score by 6.67 % , GM by 6.46 % , and MCC by 6.46 % , for KL class.
However, Table 3 shows that the increase in the number of network traffic samples increased the data size of the feature set used for the training of the S-DRNN model by 146.80 % , compared to the data size of the feature set used for the training of the DRNN model. The time taken to train the S-DRNN model with 6813554 network traffic samples in its training set was 603.83 s, and the time taken to test the model with 1,467,410 samples in its testing set was 0.7558 s. Therefore, more time was spent on training the S-DRNN model than the DRNN model.

5.3. Results of LS-DRNN Model

LAE method was introduced to address the challenge of large memory space required for the storage of the network traffic data that was used to train the S-DRNN model, as previously discussed in Section 5.2. This method reduced the feature dimensionality of 3,668,045 network traffic samples in the training set from 37 to 10 in 204.67 s. Consequently, Table 3 shows that the data size of the feature set of the LS-DRNN model reduced by 86.49 % and 94.53 % , compared to the data size of the feature sets that were used to train DRNN and S-DRNN models, respectively. Figure 7 shows that the reconstruction loss of the LAE method decreased from 0.0333 to 0.0009 as the number of epochs increased from 1 to 20. Similarly, its validation loss decreased from 0.0085 to 0.0037 . Low reconstruction loss implies that the output of LAE (i.e., low-dimensional network traffic features) is similar to the high-dimensional network traffic features without any significant loss of information.
SMOTE method generated a total of 4,612,442 synthetic network traffic samples in 15.37 s to achieve class balance across the 11 classes. Table 3 shows that the LAE method that was introduced in the LS-DRNN model reduced the sampling time by 97.20 % , compared to the sampling time in the S-DRNN model. Figure 5 shows that the training loss of the LS-DRNN model decreased from 0.1860 to 0.0068 as the number of epochs increased from 1 to 20. Similarly, its validation loss decreased from 0.1313 to 0.0128 . The consistent decrease in training loss and validation loss, as the number of epochs increased, implies that the LS-DRNN model neither under-fit nor over-fit the network traffic data in the training and validation sets, respectively.
Table 4 and Table 7 shows that the classification performance of the LS-DRNN model was better than that of the DRNN model, and it is similar to that of the S-DRNN model. Specifically, the LAE and SMOTE methods that were introduced in the training of the LS-DRNN model increased the:
  • recall by 14.01 % , F1 score by 2.65 % , BACC by 7 % , and MCC by 2.74 % , for DDH class;
  • recall by 5.71 % , and BACC by 2.86 % , for DH class;
  • recall by 2.35 % , and BACC by 1.18 % , for Norm class;
  • precision by 100 % , recall by 100 % , F1 score by 100 % , BACC by 50 % , GM by 100 % , and MCC by 100 % , for DE class; and
  • precision by 12.50 % , F1 score by 6.67 % , GM by 6.46 % , and MCC by 6.46 % , for KL class.
The detail of the detection rates and misclassification rates of the LS-DRNN model in each of the 11 classes is available in Figure 8. Table 8 and Table 9 show that the LS-DRNN model demonstrated better classification performance and faster detection speed than state-of-the-art ML and DL models. The time taken to train the LS-DRNN model with 6,813,554 network traffic samples in its training set was 616.96 s, and the time taken to test the model with 1,467,410 samples in its testing set was 0.7181 s. Table 3 shows that having the lowest testing time, the LS-DRNN model achieved the fastest detection speed, compared to DRNN and S-DRNN models. In summary, the results of this study show that the LAE method reduced the dimensionality of network traffic features in the training set from 37 to 10, and this consequently reduced the memory space required for data storage by 86.49 % . SMOTE method helped the LS-DRNN model to achieve high classification performance in minority classes, and the overall detection rate increased by 10.94 % .

6. Conclusions

In this paper, we developed a memory-efficient DL method, named LS-DRNN, to detect botnet attacks in IoT networks. This method exploited the combined benefits of an unsupervised DL method (LAE), a sampling method (SMOTE), and a supervised DL method (DRNN) for efficient performance. The effectiveness of the LS-DRNN method was evaluated with the highly imbalanced network traffic features and their corresponding labels in the Bot-IoT dataset. First, the LAE method reduced the dimensionality of the network traffic features to minimise the memory space required for the storage of training data in IoT back-end server or cloud platform. Then, SMOTE method generated a total of 4,612,442 synthetic network traffic samples to achieve class balance across the 11 classes. Lastly, DRNN performed a multi-class classification of balanced, low-dimensional network traffic features to identify botnet attacks in IoT networks.
The results of our experiments showed that the dimensionality of the network traffic features reduced from 37 to 10. Consequently, the memory space required for the storage of training data reduced by 86.49 % . Furthermore, the LS-DRNN model achieved high classification performance in the minority classes. Compared to the DRNN model, the overall precision, recall, F1 score, BACC, GM, and MCC of the LS-DRNN model increased by 8.29 % , 10.94 % , 9.59 % , 5.45 % , 8.66 % , and 9.56 % , respectively. Furthermore, the LS-DRNN model demonstrated better classification performance and faster detection speed than state-of-the-art ML and DL models. Therefore, the findings of this study validate the effectiveness of the LS-DRNN model for efficient botnet attack detection in IoT networks.

Author Contributions

Conceptualization, S.I.P., B.A. and R.A.; methodology, S.I.P., B.A., R.A., M.H., A.A.A.; software, S.I.P.; validation, S.I.P., B.A. and R.A.; formal analysis, S.I.P. and B.A.; investigation, S.I.P.; resources, B.A. and R.A.; data curation, S.I.P.; writing—original draft preparation, S.I.P.; writing—review and editing, B.A., R.A., M.H. and A.A.A.; visualization, S.I.P. and B.A.; supervision, B.A. and M.H.; project administration, B.A., R.A. and M.H.; funding acquisition, B.A. and A.A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by Cyraatek Ltd UK, the Faculty of Science & Engineering, Manchester Metropolitan University, and in part by ENERGY-IQ project, a UK-Canada Power Forward Smart Grid Demonstrator project funded by The Department for Business, Energy and Industrial Strategy (BEIS) under Grant 7454460; the NICE (Nigerian Intelligent Clean Energy) Marketplace project funded by the Department for International Development (DFID); and Covenant University Centre for Research, Innovation and Discovery (CUCRID).

Data Availability Statement

Bot-IoT data set is available from https://ieee-dataport.org/documents/bot-iot-dataset (accessed on 1 February 2021).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Aceto, G.; Persico, V.; Pescape, A. A survey on information and communication technologies for Industry 4.0: State-of-the-art, taxonomies, perspectives, and challenges. IEEE Commun. Surv. Tutor. 2019, 21, 3467–3501. [Google Scholar] [CrossRef]
  2. Williams, J.M.; Khanna, R.; Ruiz-Rosero, J.P.; Pisharody, G.; Qian, Y.; Carlson, C.R.; Liu, H.; Ramirez-Gonzalez, G. Weaving the wireless web: Toward a low-power, dense wireless sensor network for the industrial IoT. IEEE Microw. Mag. 2017, 18, 40–63. [Google Scholar] [CrossRef]
  3. Darvishi, H.; Ciuonzo, D.; Eide, E.R.; Rossi, P.S. Sensor-Fault Detection, Isolation and Accommodation for Digital Twins via Modular Data-Driven Architecture. IEEE Sens. J. 2020. [Google Scholar] [CrossRef]
  4. Vormayr, G.; Zseby, T.; Fabini, J. Botnet communication patterns. IEEE Commun. Surv. Tutor. 2017, 19, 2768–2796. [Google Scholar] [CrossRef]
  5. HaddadPajouh, H.; Dehghantanha, A.; Parizi, R.M.; Aledhari, M.; Karimipour, H. A survey on internet of things security: Requirements, challenges, and solutions. Internet Things 2019, 100129. [Google Scholar] [CrossRef]
  6. Tange, K.; De Donno, M.; Fafoutis, X.; Dragoni, N. A systematic survey of industrial Internet of Things security: Requirements and fog computing opportunities. IEEE Commun. Surv. Tutor. 2020, 22, 2489–2520. [Google Scholar] [CrossRef]
  7. Farooq, M.J.; Zhu, Q. Modeling, analysis, and mitigation of dynamic botnet formation in wireless iot networks. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2412–2426. [Google Scholar] [CrossRef] [Green Version]
  8. Zhao, B.; Ji, S.; Lee, W.H.; Lin, C.; Weng, H.; Wu, J.; Zhou, P.; Fang, L.; Beyah, R. A Large-scale Empirical Study on the Vulnerability of Deployed IoT Devices. IEEE Trans. Dependable Secur. Comput. 2020, 1. [Google Scholar] [CrossRef]
  9. Yin, L.; Luo, X.; Zhu, C.; Wang, L.; Xu, Z.; Lu, H. ConnSpoiler: Disrupting C&C communication of IoT-based botnet through fast detection of anomalous domain queries. IEEE Trans. Ind. Inform. 2019, 16, 1373–1384. [Google Scholar]
  10. Popoola, S.I.; Adebisi, B.; Hammoudeh, M.; Gui, G.; Gacanin, H. Hybrid Deep Learning for Botnet Attack Detection in the Internet-of-Things Networks. IEEE Internet Things J. 2021, 8, 4944–4956. [Google Scholar] [CrossRef]
  11. Koroniotis, N.; Moustafa, N.; Sitnikova, E.; Turnbull, B. Towards the development of realistic botnet dataset in the internet of things for network forensic analytics: Bot-iot dataset. Future Gener. Comput. Syst. 2019, 100, 779–796. [Google Scholar] [CrossRef] [Green Version]
  12. Yu, J. HEH, a New IoT P2P Botnet Going after Weak Telnet Services, October 2020. Available online: https://blog.netlab.360.com/heh-an-iot-p2p-botnet/ (accessed on 19 November 2020).
  13. Raza, S.; Wallgren, L.; Voigt, T. SVELTE: Real-time intrusion detection in the Internet of Things. AD Hoc Netw. 2013, 11, 2661–2674. [Google Scholar] [CrossRef]
  14. Bertino, E.; Islam, N. Botnets and internet of things security. Computer 2017, 50, 76–79. [Google Scholar] [CrossRef]
  15. Mishra, P.; Varadharajan, V.; Tupakula, U.; Pilli, E.S. A detailed investigation and analysis of using machine learning techniques for intrusion detection. IEEE Commun. Surv. Tutor. 2018, 21, 686–728. [Google Scholar] [CrossRef]
  16. Buczak, A.L.; Guven, E. A survey of data mining and machine learning methods for cyber security intrusion detection. IEEE Commun. Surv. Tutor. 2015, 18, 1153–1176. [Google Scholar] [CrossRef]
  17. Xin, Y.; Kong, L.; Liu, Z.; Chen, Y.; Li, Y.; Zhu, H.; Gao, M.; Hou, H.; Wang, C. Machine learning and deep learning methods for cybersecurity. IEEE Access 2018, 6, 35365–35381. [Google Scholar] [CrossRef]
  18. Gümüşbaş, D.; Yıldırım, T.; Genovese, A.; Scotti, F. A Comprehensive Survey of Databases and Deep Learning Methods for Cybersecurity and Intrusion Detection Systems. IEEE Syst. J. 2020. [Google Scholar] [CrossRef]
  19. Chaabouni, N.; Mosbah, M.; Zemmari, A.; Sauvignac, C.; Faruki, P. Network intrusion detection for IoT security based on learning techniques. IEEE Commun. Surv. Tutor. 2019, 21, 2671–2701. [Google Scholar] [CrossRef]
  20. Mahdavifar, S.; Ghorbani, A.A. Application of deep learning to cybersecurity: A survey. Neurocomputing 2019, 347, 149–176. [Google Scholar] [CrossRef]
  21. Ferrag, M.A.; Maglaras, L. DeepCoin: A Novel Deep Learning and Blockchain-Based Energy Exchange Framework for Smart Grids. IEEE Trans. Eng. Manag. 2019. [Google Scholar] [CrossRef] [Green Version]
  22. Ferrag, M.A.; Maglaras, L.; Moschoyiannis, S.; Janicke, H. Deep learning for cyber security intrusion detection: Approaches, datasets, and comparative study. J. Inf. Secur. Appl. 2020, 50, 102419. [Google Scholar] [CrossRef]
  23. Alkadi, O.; Moustafa, N.; Turnbull, B.; Choo, K.K.R. A Deep Blockchain Framework-enabled Collaborative Intrusion Detection for Protecting IoT and Cloud Networks. IEEE Internet Things J. 2020. [Google Scholar] [CrossRef]
  24. Lee, S.; Abdullah, A.; Jhanjhi, N.; Kok, S. Classification of botnet attacks in IoT smart factory using honeypot combined with machine learning. PeerJ Comput. Sci. 2021, 7, e350. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, Y.; Xu, J.; Wang, Z.; Geng, R.; Choo, K.K.R.; Pérez-Díaz, J.A.; Zhu, D. Efficient and Intelligent Attack Detection in Software Defined IoT Networks. In Proceedings of the 2020 IEEE International Conference on Embedded Software and Systems (ICESS), Shanghai, China, 10–11 December 2020; pp. 1–9. [Google Scholar]
  26. Filus, K.; Domańska, J.; Gelenbe, E. Random Neural Network for Lightweight Attack Detection in the IoT. In Symposium on Modelling, Analysis, and Simulation of Computer and Telecommunication Systems; Springer: Berlin, Germany, 2020; pp. 79–91. [Google Scholar]
  27. Lawal, M.A.; Shaikh, R.A.; Hassan, S.R. An Anomaly Mitigation Framework for IoT Using Fog Computing. Electronics 2020, 9, 1565. [Google Scholar] [CrossRef]
  28. Oreški, D.; Andročec, D. Genetic algorithm and artificial neural network for network forensic analytics. In Proceedings of the 2020 43rd International Convention on Information, Communication and Electronic Technology (MIPRO), Opatija, Croatia, 28 September–2 October 2020; pp. 1200–1205. [Google Scholar]
  29. NG, B.A.; Selvakumar, S. Anomaly detection framework for Internet of things traffic using vector convolutional deep learning approach in fog environment. Future Gener. Comput. Syst. 2020, 113, 255–265. [Google Scholar]
  30. Ibitoye, O.; Shafiq, O.; Matrawy, A. Analyzing adversarial attacks against deep learning for intrusion detection in IoT networks. arXiv 2019, arXiv:1905.05137. [Google Scholar]
  31. Kumar, P.; Gupta, G.P.; Tripathi, R. Toward Design of an Intelligent Cyber Attack Detection System using Hybrid Feature Reduced Approach for IoT Networks. Arab. J. Sci. Eng. 2021, 46, 3749–3778. [Google Scholar] [CrossRef]
  32. Kumar, P.; Kumar, R.; Gupta, G.P.; Tripathi, R. A Distributed framework for detecting DDoS attacks in smart contract-based Blockchain-IoT Systems by leveraging Fog computing. Trans. Emerg. Telecommun. Technol. 2020, e4112. [Google Scholar] [CrossRef]
  33. Shafiq, M.; Tian, Z.; Bashir, A.K.; Du, X.; Guizani, M. Corrauc: A malicious bot-iot traffic detection method in iot network using machine learning techniques. IEEE Internet Things J. 2020. [Google Scholar] [CrossRef]
  34. Koroniotis, N.; Moustafa, N.; Sitnikova, E. A new network forensic framework based on deep learning for Internet of Things networks: A particle deep framework. Future Gener. Comput. Syst. 2020, 110, 91–106. [Google Scholar] [CrossRef]
  35. Shafiq, M.; Tian, Z.; Bashir, A.K.; Du, X.; Guizani, M. IoT malicious traffic identification using wrapper-based feature selection mechanisms. Comput. Secur. 2020, 94, 101863. [Google Scholar] [CrossRef]
  36. Asadi, M.; Jamali, M.A.J.; Parsa, S.; Majidnezhad, V. Detecting botnet by using particle swarm optimization algorithm based on voting system. Future Gener. Comput. Syst. 2020, 107, 95–111. [Google Scholar] [CrossRef]
  37. Huong, T.T.; Bac, T.P.; Long, D.M.; Thang, B.D.; Luong, T.D.; Binh, N.T. An Efficient Low Complexity Edge-Cloud Framework for Security in IoT Networks. In Proceedings of the 2020 IEEE Eighth International Conference on Communications and Electronics (ICCE), Phu Quoc Island, Vietnam, 13–15 January 2021; pp. 533–539. [Google Scholar]
  38. Kumar, P.; Gupta, G.P.; Tripathi, R. TP2SF: A Trustworthy Privacy-Preserving Secured Framework for sustainable smart cities by leveraging blockchain and machine learning. J. Syst. Archit. 2020, 115, 101954. [Google Scholar] [CrossRef]
  39. Alshamkhany, M.; Alshamkhany, W.; Mansour, M.; Khan, M.; Dhou, S.; Aloul, F. Botnet Attack Detection using Machine Learning. In Proceedings of the 2020 14th International Conference on Innovations in Information Technology (IIT), Al Ain, United Arab Emirates, 17–18 November 2020; pp. 203–208. [Google Scholar]
  40. Sriram, S.; Vinayakumar, R.; Alazab, M.; Soman, K. Network Flow based IoT Botnet Attack Detection using Deep Learning. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Toronto, ON, Canada, 6–9 July 2020; pp. 189–194. [Google Scholar]
  41. Liaqat, S.; Akhunzada, A.; Shaikh, F.S.; Giannetsos, A.; Jan, M.A. SDN orchestration to combat evolving cyber threats in Internet of Medical Things (IoMT). Comput. Commun. 2020, 160, 697–705. [Google Scholar] [CrossRef]
  42. Soe, Y.N.; Santosa, P.I.; Hartanto, R. DDoS Attack Detection Based on Simple ANN with SMOTE for IoT Environment. In Proceedings of the 2019 Fourth International Conference on Informatics and Computing (ICIC), Semarang, Indonesia, 16–17 October 2019; pp. 1–5. [Google Scholar] [CrossRef]
  43. Bagui, S.; Li, K. Resampling imbalanced data for network intrusion detection datasets. J. Big Data 2021, 8, 1–41. [Google Scholar] [CrossRef]
  44. Derhab, A.; Aldweesh, A.; Emam, A.Z.; Khan, F.A. Intrusion Detection System for Internet of Things Based on Temporal Convolution Neural Network and Efficient Feature Engineering. Wirel. Commun. Mob. Comput. 2020, 2020, 6689134. [Google Scholar] [CrossRef]
  45. Popoola, S.I.; Adebisi, B.; Ande, R.; Hammoudeh, M.; Anoh, K.; Atayero, A.A. SMOTE-DRNN: A Deep Learning Algorithm for Botnet Detection in the Internet-of-Things Networks. Sensors 2021, 21, 2985. [Google Scholar] [CrossRef]
  46. Mulyanto, M.; Faisal, M.; Prakosa, S.W.; Leu, J.S. Effectiveness of Focal Loss for Minority Classification in Network Intrusion Detection Systems. Symmetry 2021, 13, 4. [Google Scholar] [CrossRef]
  47. Injadat, M.; Moubayed, A.; Shami, A. Detecting botnet attacks in IoT environments: An optimized machine learning approach. arXiv 2020, arXiv:2012.11325. [Google Scholar]
  48. Khan, M.A.; Kim, Y. Deep Learning-Based Hybrid Intelligent Intrusion Detection System. Comput. Mater. Contin. 2021, 68, 671–687. [Google Scholar] [CrossRef]
  49. Roopak, M.; Yun Tian, G.; Chambers, J. Deep Learning Models for Cyber Security in IoT Networks. In Proceedings of the 2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 7–9 January 2019; pp. 0452–0457. [Google Scholar] [CrossRef]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  51. Ma, X.; Shi, W. AESMOTE: Adversarial Reinforcement Learning with SMOTE for Anomaly Detection. IEEE Trans. Netw. Sci. Eng. 2020. [Google Scholar] [CrossRef]
  52. Ling, C.X.; Li, C. Data Mining for Direct Marketing: Problems and Solutions; Kdd: Washington, DC, USA, 1998; Volume 98, pp. 73–79. [Google Scholar]
  53. Japkowicz, N. The class imbalance problem: Significance and strategies. In Proceedings of the International Conference on Artificial Intelligence; Citeseer: University Park, PA, USA, 2000; Volume 56. [Google Scholar]
  54. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  55. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  56. Ferrag, M.A.; Maglaras, L.; Ahmim, A.; Derdour, M.; Janicke, H. RDTIDS: Rules and Decision Tree-Based Intrusion Detection System for Internet-of-Things Networks. Future Internet 2020, 12, 44. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Block diagram showing the proposed methodology.
Figure 1. Block diagram showing the proposed methodology.
Electronics 10 01104 g001
Figure 2. LAE model for feature dimensionality reduction.
Figure 2. LAE model for feature dimensionality reduction.
Electronics 10 01104 g002
Figure 3. SMOTE method for the generation of synthetic network traffic features.
Figure 3. SMOTE method for the generation of synthetic network traffic features.
Electronics 10 01104 g003
Figure 4. DRNN model for multi-class classification.
Figure 4. DRNN model for multi-class classification.
Electronics 10 01104 g004
Figure 5. Cross-entropy losses during training.
Figure 5. Cross-entropy losses during training.
Electronics 10 01104 g005
Figure 6. Cross-entropy losses during validation.
Figure 6. Cross-entropy losses during validation.
Electronics 10 01104 g006
Figure 7. Reconstruction losses of LAE model.
Figure 7. Reconstruction losses of LAE model.
Electronics 10 01104 g007
Figure 8. Confusion matrix of LS-DRNN model.
Figure 8. Confusion matrix of LS-DRNN model.
Electronics 10 01104 g008
Table 1. Review of related works.
Table 1. Review of related works.
Dimensionality Reduction MethodClass Balance MethodReferences
Feature selectionNone[11,24,25,26,27,28,29,30,31,32,33,34,35,36]
PCANone[37,38,39]
t-SNENone[40]
LAENone[10]
NoneUp-sampling[41]
NoneSMOTE[42,43,44,45]
Feature selectionFocal loss[46]
Feature selectionSMOTE[47]
LAESMOTEThis paper
Table 2. Network traffic samples in the training, validation, and testing sets.
Table 2. Network traffic samples in the training, validation, and testing sets.
ClassTrainingValidationTesting
DDH588197204
DDT586,393195,713195,274
DDU568,760189,407190,088
DH906311268
DT369,965122,861122,974
DU619,414206,772206,789
Norm29086101
OSF10,79535373582
SS43,94914,80614,413
DE411
KL481411
Table 3. Data size and computation time.
Table 3. Data size and computation time.
ModelData Size (MB)Encoding Time (s)Sampling Time (s)Train Time (s)Test Time (s)
Ferrag et al. [21]1085.88----
Ferrag et al. [22]1085.88----
Ferrag et al. [56]1085.88----
Alkadi et al. [23]1085.88----
DRNN1085.88--207.720.7552
S-DRNN2680.24-548.72603.830.7558
LS-DRNN146.74204.6715.37616.960.7181
Table 4. Classification performance of DRNN model.
Table 4. Classification performance of DRNN model.
MetricDDHDDTDDUDHDTDUNormOSFSSDEKLOverall
Accuracy99.9999.94100.0099.9999.95100.00100.00100.00100.00100.00100.0099.99
Precision95.0499.7999.9992.0999.99100.00100.00100.0099.980.0087.5088.58
Recall85.35100.00100.0094.2999.6899.9997.65100.00100.000.00100.0088.81
F1 score89.9399.8999.9993.1899.8499.9998.81100.0099.990.0093.3388.63
FPR0.000.080.000.000.000.000.000.000.000.000.000.01
NPV100.00100.00100.00100.0099.9499.99100.00100.00100.00100.00100.0099.99
BACC92.6799.96100.0097.1499.8499.9998.82100.00100.0050.00100.0094.40
GM97.4899.8999.9995.9699.96100.00100.00100.0099.990.0093.5489.71
MCC90.0699.8699.9993.1899.8099.9998.82100.0099.990.0093.5488.66
Table 5. Sample distribution of training set for S-DRNN model.
Table 5. Sample distribution of training set for S-DRNN model.
ClassOriginal DataAfter SamplingGenerated Samples
DDH588619,414618,826
DDT586,393619,41433,021
DDU568,760619,41450,654
DH906619,414618,508
DT369,965619,414249,449
DU619,414619,4140
Norm290619,414619,124
OSF10,795619,414608,619
SS43,949619,414575,465
DE4619,414619,410
KL48619,414619,366
Total2,201,1126,813,5544,612,442
Table 6. Classification performance of S-DRNN model.
Table 6. Classification performance of S-DRNN model.
MetricDDHDDTDDUDHDTDUNormOSFSSDEKLOverall
Accuracy100.0099.99100.00100.0099.99100.00100.00100.00100.00100.00100.00100.00
Precision100.0099.99100.0099.5299.94100.00100.00100.0099.99100.00100.0099.95
Recall100.0099.97100.0099.5299.99100.00100.0099.96100.00100.00100.0099.95
F1 score100.0099.98100.0099.5299.97100.00100.0099.98100.00100.00100.0099.95
FPR0.000.000.000.000.010.000.000.000.000.000.000.00
NPV100.0099.99100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00
BACC100.0099.98100.0099.7699.99100.00100.0099.98100.00100.00100.0099.97
GM100.0099.99100.0099.7699.97100.00100.00100.00100.00100.00100.0099.97
MCC100.0099.97100.0099.5299.96100.00100.0099.98100.00100.00100.0099.95
Table 7. Classification performance of LS-DRNN model.
Table 7. Classification performance of LS-DRNN model.
MetricDDHDDTDDUDHDTDUNormOSFSSDEKLOverall
Accuracy100.0099.6899.9399.9999.6999.93100.00100.00100.00100.00100.0099.93
Precision86.6799.1699.9587.1499.5199.7993.41100.0099.98100.00100.0096.87
Recall99.3699.6699.77100.0098.6599.95100.0099.9399.94100.00100.0099.75
F1 score92.5899.4199.8693.1399.0899.8796.5999.9699.96100.00100.0098.22
FPR0.000.310.020.010.100.080.000.000.000.000.000.05
NPV100.0099.8899.92100.0099.7399.98100.00100.00100.00100.00100.0099.95
BACC99.6899.6799.87100.0099.2899.93100.0099.9699.97100.00100.0099.85
GM93.0999.5299.9393.3599.6299.8896.65100.0099.99100.00100.0098.37
MCC92.8099.1999.8193.3498.8999.8296.6599.9699.96100.00100.0098.22
Table 8. Recall of ML and DL models across 11 classes.
Table 8. Recall of ML and DL models across 11 classes.
ModelDDHDDTDDUDHDTDUNormOSFSSDEKL
RNN [21]100.00100.00100.00100.00100.00100.00-92.2287.9199.7577.91
SVM [21]62.2489.5698.1470.1471.26100.00-70.1472.8289.6765.12
RF [21]82.2688.2855.2682.2081.7782.99-82.2069.8286.5581.56
NB [21]50.7878.6778.5068.6865.56100-68.6865.2166.5565.62
DNN [22]96.6296.2296.1296.7096.6396.53-96.1496.43100.0096.76
RNN [22]96.5696.6596.6796.8796.7796.76-96.7696.87100.0097.00
CNN [22]97.0197.0097.0197.5197.1197.11-97.0097.10100.0098.10
RBM [22]96.5496.5196.5296.8096.5796.56-96.3096.30100.0097.11
DBN [22]96.7296.6096.6296.9196.7296.83-96.6196.60100.0097.66
DBM [22]96.2196.0896.1196.9996.3396.65-96.0896.07100.0098.22
DAE [22]97.9997.7197.9998.4198.0098.03-97.7297.71100.0098.33
RDTIDS [56]93.1795.8498.6677.47100.00100.00-98.1699.47100.00100.00
BLSTM [23]99.2599.1099.4599.7599.6599.79-92.7792.2096.5089.90
LS-DRNN99.3699.6699.77100.0098.6599.95100.0099.9399.94100.00100.00
Table 9. Overall classification performance of ML and DL models.
Table 9. Overall classification performance of ML and DL models.
ModelAccuracyPrecisionRecallF1 ScoreFPRNPVAUCGMMCC T train (s) T test (s)
RNN [21]99.91---1.28----201.7044.23
SVM [21]----2.99------
RF [21]----4.29------
NB [21]----3.24------
DNN [22]98.22---1.14----991.60-
RNN [22]98.31---1.10----1400.60-
CNN [22]98.37---1.00----1367.20-
RBM [22]98.28---1.13----2111.90-
DBN [22]98.31---1.12----2921.70-
DBM [22]98.38---1.11----2800.10-
DAE [22]98.39---1.11----2816.20-
RDTIDS [56]97.00---1.12----195.502.27
BLSTM [23]98.91---1.20----149.6069.10
LS-DRNN99.9396.8799.7598.220.0599.9599.8599.3798.22616.960.72
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Popoola, S.I.; Adebisi, B.; Ande, R.; Hammoudeh, M.; Atayero, A.A. Memory-Efficient Deep Learning for Botnet Attack Detection in IoT Networks. Electronics 2021, 10, 1104. https://doi.org/10.3390/electronics10091104

AMA Style

Popoola SI, Adebisi B, Ande R, Hammoudeh M, Atayero AA. Memory-Efficient Deep Learning for Botnet Attack Detection in IoT Networks. Electronics. 2021; 10(9):1104. https://doi.org/10.3390/electronics10091104

Chicago/Turabian Style

Popoola, Segun I., Bamidele Adebisi, Ruth Ande, Mohammad Hammoudeh, and Aderemi A. Atayero. 2021. "Memory-Efficient Deep Learning for Botnet Attack Detection in IoT Networks" Electronics 10, no. 9: 1104. https://doi.org/10.3390/electronics10091104

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop