Introduction

One of the major and fundamental issue is emotion recognition during the development of an interactive computer system [1,2,3]. Recognition of facial emotion/expression is essential, because nowadays it place its wide applications in various sectors like psychological distress and pain detection [4]. Some fields like psychology, sociology, and automatic expression recognition, therefore provided a considerable importance for this emotion recognition process to create a highly user affable software and user agents in these fields. This process of automatic facial expression recognition (FER) has exhibited its large implications in the human computer interaction (HCI) field [5]. Recently, the affective computing is considered as the most significant study field in HCI, which highly intends to improve the human–machine interaction by clearly recognizing the emotion states from the human faces. The study required for the ER states has been analyzed from various information sources like speech signals [6], biological signals [7], images and video sequences [8, 9]. In this proposed method, the expressions required for ER are gathered from video sequence. Number of experimental analysis are carried out in this automatic FER, as it implies huge practical importance in driver fatigue surveillance, medical treatment, sociable robotics, and also in few HCI systems. Through the field of machine learning and computer vision, a number of FER systems are largely explored to gather the information regarding the expressions from the facial representations [10].

Researchers have analysed number of FER techniques having highest accuracy, but still it remains a challenging task of capturing the emotions from the faces of different nationalities and different angles [11]. Facial expression is identified as the most beneficiary source for capturing the human behavior and attitude [12]. A report provided by psychological study illustrates that the features found around the eyes, mouth, and nose are highly essential for FER purpose. In various literary works, number of different techniques have been developed for the purpose of FER in still images and video data, whereas for classification purpose different types of classifiers like decision tree (DT), Support Vector Machine (SVM), Bayesian network, neural networks (NN), K-Nearest Neighbour (KNN), and rule-based classifiers are used [13]. The ultimate aim of this classifiers is to give a perfect trade-off between the recognition accuracy and processing time [14]. Technically, this ER process is categorized into two different groups they are (i) Individual ER, (ii) Group-level ER. However, the studies found in individual based ER are found mature, but the research process in this group level ER is still in its early stages. Major challenges exhibited by this group level ER is the detection of faces from group and the aggregation of emotions content from this group through video or image [15].

Group based ER (GER) [16] is gaining a huge attention in the affective computing field. This GER is largely required for various applications like improvement in e-classroom teaching on the basis of students face expression, analyzing the crowd behavior, advertisements, typical photos best representing events, and so on. Rapid improvement has been encountered in this field due to recently occurred challenges such as EmotiW [17, 18]. Major challenging task of this classification process is to given the group image into Happy, Sad, Angry and Fear. This classification process is accurately achieved by extracting the emotions details from all the faces in the image [19, 20]. GER is now grown as a trending topic because the demands for recognizing the emotion from the group of people are gaining a maximum interest in both the social media and security arena. The World Wide Web (WWW) and multimedia are advanced a lot, therefore the users upload the number of videos and images in the internet over the social networking platforms. Due to the difference in indoor and outdoor settings, the recognition of emotions from the faces of multiple persons becomes somewhat a difficult task. Therefore, in this method even more challenging task is considered which is to classify the emotions that are shared by a group of people as Happy, Sad, Angry and Fear. This classification process is normally referred as group-level emotion recognition. While compared to other emotion recognition process, this GER is found more challenging as it depends on single image (lack of temporal information), moreover the human faces are often occurring at low resolution (lack of facial details).

In this paper we introduced the RFNN with SSD optimization algorithm for the classification purpose. The advantages of SSD is it reduces the complexity and increases the convergence speed, thus we use SSD in RFNN to get the better accuracy performance. As well as here we use modified sea lion optimization algorithm for feature selection. The existing papers does not have any feature selection techniques for GER which reduces the accuracy. Modified sea lion optimization is the combination of sea lion optimization with GSO optimization algorithm. For finding the best features, first we update the position by using GSO algorithm. GSO position is updated in sea lion optimization to get the optimal feature. The ultimate aim of this technique is to improve a system that can automatically classify the emotions from the group of peoples as Happy, Sad, Angry and Fear. Few other objectives of this process are:

  • To correctly detect the face from crowded image.

  • To design a technique that can effectively classify the emotion with highest accuracy and less error value.

  • To reduce the dimension of extracted features using a novel adaptive algorithm.

  • To improve the accuracy with less misclassification error.

The rest of the paper is organized as follows: “Related work” section gives some analysis of related work. Pre-processing, feature extraction, feature selection and classification is briefly discussed in “Proposed method for group based emotion recognition” section. The performance analysis of proposed method is explained in results and discussion which is in “Results and discussion” section. The conclusion of the work is present in “Conclusion” section.

Related work

For image based group emotion recognition, the influence of mid-level and low-level features was investigated by Balaji et al. [21]. They hypothesize that the objects and faces of humans are the major information sources and therefore it can serve as the mid-level features. Hence, using pre-trained Deep Net models they detect the objects and faces. With different encoding methods the information from different layers was extensively investigated for attain the richest feature vectors. To yield the best feature, vector of locally aggregated descriptors (VLAD) and Fisher vector encoding schemes were investigated.

Surace et al. [22] implemented a solution based on a new combination of Bayesian classifier and deep neural networks. The deep learning methods works on bottom-up techniques which evaluating emotions conveyed by isolated faces. A global emotion integrating top down features was estimated by Bayesian classifier which obtained by a scene descriptor. They tested the framework on the dataset in the wild challenge 2017 for the emotion recognition to validate the system.

In the category of group level emotion recognition Abbas et al. [23] presented the proposed solution for the emotion recognition. To classify a group of people emotion as positive, negative and neutral was one of the main objectives. For classification, their method incorporates both facial and image context information which extracted from an image. To predict facial emotions in an image from detected faces, they utilized convolutional neural networks (CNN). In order to perform group level emotion recognition, different techniques are explored by training and combining these two deep neural networks.

In an image, an emotional detection for a group of people was described by Shamsi et al. [24]. They utilized the combination of both top-down and bottom-up method. Using face-level predictions they create heat maps and train ConvNets on them. By constructing heat maps for emotions intensity they perform face level emotion detection and join individual images. At last, an ensemble of both top down and bottom up methods was created.

GER is considered as a challenging issue, due to some obstructions like image quality, actor’s variance, variation in head and body pose, different outdoor and indoor settings, and occlusions. Less number of research works are carried out in this GER process, but now it is gaining high popularity. This is due to the availability of large amount of data in social networks, most particularly the images of group of people those who are participating in social events. Most essential clue for identifying the individual behaviour is facial expression analysis, but this analysis process is found challenging due to the existence of few variations in illuminations, different facial tones, and face poses. A large number of techniques are introduced in existing techniques for emotion recognition, but none of them attains required accuracy. Thus, we introduced the RFNN with SSD optimization algorithm for classification purpose. The advantages of SSD is it reduces the complexity and increases the convergence speed, thus we use SSD in RFNN to get the better accuracy performance. As well as here we use modified sea lion optimization algorithm for feature selection. In previous works, they did not use the feature selection technique for group based emotion recognition. Feature selection techniques are utilized here, thus will select best features and it will remove to add more number of unwanted features.

Proposed method for group based emotion recognition

Figure 1 shows the basic work flow for facial emotion recognition. At first, a group of images showing different emotions like Happy, Sad, Angry and Fear are applied in this method for pre-processing step. Additionally, in this pre-processing step both illumination and normalization are included. Then, the MLTP, GLCM and LESH are applied for feature extraction approach. These three techniques are applied from the pre-processed image to extract the features. An optimal feature selection process is done by modified sea lion optimization algorithm. Then, with this optimal features the emotion recognition is done by RFNN classifier using SSD.

Fig. 1
figure 1

Proposed work flow

In our work, emotion recognition from group of people is performed by utilizing the Neural Network. From the multi user video data, initially, original video frames are taken as input and pre-process it. During pre-processing, the input image is denoised by Fast averaging peer group (FAPG) filter. The face from that pre-processed image is detected using Viola Jones classifier. Then, an algorithm named as eye-center detection is introduced along with this face detection algorithm to improve the recognition result. The features from this detected face is extracted for recognition purpose. Here, the MLTP, GLCM and LESH based features are extracted and carried out as the features in the further processing steps. After extracting the features, certain features are selected using Modified Sea-lion optimization algorithm process. Finally, recurrent fuzzy neural network (RFNN) classifier using Social Ski-Driver (SSD), optimization algorithm is proposed for updating the weights in the RFNN.

Pre-processing

The process of pre-processing is it suppresses unwanted distortions and improve the image data for further processing. Initially, the face image is given as an input for FAPG to remove the noise and also to control the redundant information. The features of each frames gets enhanced because of this pre-processing.

Fast averaging peer group filter

In digital images, noise reduction is considered as the problematic task [25]. The corrupted images are colour images by means of impulsive noise. To overcome this problem, the fast denoising FAPG filter is used. This filter can restore the images while maintaining the tiny images and edges also it outperforms on the high noise contamination levels. Consider the colour image as the two-dimensional array, which consists of N amount of pixels \(X_{j} = \left( {X_{j1} ,\;X_{j2} ,\;X_{j3} } \right)\) through the index value (j = 1,…N) which indicates the pixel position on the image domain. For q = 1, 2, 3 the vector components \(X_{jq} \in \left[ {0,1} \right]\) denote the colour channel values computed into integer domain in a given colour space. Also, the indexes are assigned to pixels that belongs to the local filtering window W, whereas \(X_{1}\) indicates the central pixel and \(X_{2} , \ldots X_{n}\) represents the neighbouring pixels and n signifies the size of window.

The two main steps are associated with this filter: Peer group size determination and Pixel replacement. The peer group size or else the amount of close neighbours (CN) of central pixel \(X_{1}\) is evaluated in the first step. In a given colour space, when standardized Euclidean distance \(\rho \left( {X_{i} ,X_{1} } \right)\) is lower than the pre-defined threshold value d then a pixel \(X_{i} \ne X_{1}\) going to W is a CN of \(X_{1}\). The size of peer group is expressed as:

$$m_{k} = \, \ne \left\{ {X_{j} \in W:\left\| {X_{k} - X_{j} } \right\| < d} \right\}$$
(1)

Therefore, peer group size can be represented as a pixel distortion measure produced by noise method. Pixel is treated as corrupted when value of m is low otherwise it will not be corrupted. The next step is pixel replacement. Filtering process is done after the calculation of all m values of the image pixels. Proper selection of two filter parameters such as \(d\) and \(\gamma\) play a significant role in noise reduction. First one is threshold parameter (d) which is stated as \(0 \le d \le 1\). Hence, d = 0 means two identical pixels. The maximum Euclidean distance is d = 1 in colour space. The weights \(we_{i} ,\;\;i = 2, \ldots n\) of the pixels \(X_{i}\) can be evaluated using

$$we_{i} = \frac{{\mu_{i} }}{{\sum\nolimits_{i = 2}^{n} {\mu_{i} } }},\,\mu_{i} = m_{i}^{\gamma }$$
(2)

From the above equation, the second parameter is \(\gamma\). If \(\gamma > 0\), this affects the quality of results. The parameter \(\gamma\) gives the chance to manage the degree of membership of neighbouring pixels. The differences in neighbouring pixels the peer group sizes are reduced if \(0 < \gamma < 1\). Also, if \(\gamma > 1\) the differences in neighbouring pixels the peer group sizes are increased.

Detection by Viola-Jones and eye center detection

One of the significant step is face detection which found in practically all emotion recognition methods. The face detection based on this steps like nose, mouth, eyes etc. To form an expression these component plays a major role. Using Voila and Jones algorithm [26] the face region has been detected in the proposed algorithm. Four stages are presented in Voila-Jones algorithm: haarfeatures selection, creating an integral image, Adaboost training, and cascading classifiers. Some common features are in human faces which are extracted. For face detection, to select the Viola-Jones algorithm has some main reason which are: (1) it is very robust with very low false positive rate and very high detection rate and (2) it tends to be utilized for real-time scenes because of its proficiency of preparing 25 frames for each second. Thereafter, a novel eye center detection algorithm [27] (Algorithm 1) has been proposed to locate the eye centres in the faces. In such a way the algorithm is utilized to calculate L that it can handle the head deflection also.

$$L\, = \,\sqrt {\left( {\left( {RE\,\left( {x,Center} \right)\, - \,LE\left( {x,\,Center} \right)} \right)^{2} \, + \,\left( {RE\left( {y,\,Center} \right)\, - \,LE\,\left( {y,\,Center} \right)} \right)^{2} } \right)}$$
(3)

The input video is shown in Fig. 2 and face detection by Voila & Jones and Eye Center Detection is shown in Fig. 3.

figure a
Fig. 2
figure 2

Input video

Fig. 3
figure 3

Face detection by Voila and Jones and eye center detection

Feature extraction

The feature extraction is necessary to perform from the object detected image. This features are very useful to accurately recognize the emotion from the video. Here, three different feature extraction methods are applied they are MLTP, GLCM and LESH.

Multivariate local texture pattern

From a 3 × 3 neighbourhood, the local texture pattern model extracts local texture information. Let ‘X’ denoted by local neighbourhood. At that point the connection between centre pixel and its eight neighbours is characterized as below:

$$P\,(g_{i} \,,\,g_{c} )\, = \,\left\{ \begin{aligned} & 0\,\,\,\,if\,g_{i} \, \prec \,\left( {g_{c} \, - \,\Delta_{g} } \right) \\ & 1\,\,\,\,\,if\,\left( {g_{c} \, - \,\Delta_{g} } \right)\, \le \,g_{i} \, \le \,\left( {g_{c} \, + \,\Delta_{g} } \right) \\ & 9\,\,\,\,if\,g_{i} \, \succ \,\left( {g_{c} \, + \,\Delta_{g} } \right) \\ \end{aligned} \right.$$
(4)

where, \(\Delta_{g}\) is a small positive value with the centre pixel which is set to communicate the closeness of neighbouring pixel. In the neighbourhood, the output level assigned to the ith pixel is \(p\,\left( {g_{i} \,,\,g_{c} } \right)\). The MLTP output is shown in Fig. 4.

Fig. 4
figure 4

MLTP output

Gray-level co-occurrence matrix

We extract features from image using the GLCM. Feature extracted from GLCM are correlation, entropy, contrast, variance, average, angular second moment, dissimilarity, homogeneity, energy, mean, standard deviation and maximum probability.

Local energy based shape histogram

For the image of interest, LESH feature extraction method is primarily depended upon the idea of calculating the local energy pattern histogram (Table 1). Using phase congruency scheme, the local energies are calculated with different orientations.

Table 1 Gray-level co-occurrence matrix

Local energy is computed as follow

$$E\, = \,\frac{{\sum\nolimits_{n} {W\,(x)\,\left[ {A_{n} \,(x)\,\left( {\cos \left( {\phi_{n} \,(x) - \varphi \,(x)} \right) - \left| {\sin \,\left( {\phi_{n} \,(x) - \varphi \,(x)} \right)} \right|} \right) - T} \right]} }}{{\sum\nolimits_{n} {A_{n} \,\left( x \right)\, + \,\varepsilon } }}$$
(5)

where, \(T\) represents the noise cancellation factor, \(W\) represents the sine of phase deviation and factor.

The local histogram \(h\) is expressed as below:

$$h_{r,b} \, = \,\sum {W_{r} \, \times \,E\, \times \,\delta_{LB} }$$
(6)

where, \(\delta_{LB}\) represents the Kronecker’s delta, local energy is computed by \(E\) and \(W_{r}\) is represented as Gaussian weighting function of region \(r\), orientation label map is represented by \(L\) and current bin is represented as \(b\).

Gaussian weighting function of region \(r\) is given as follow.

$$W_{r} \, = \,\frac{1}{{\sqrt {2\pi \sigma } }}e^{{{\raise0.7ex\hbox{${\left[ {\left( {x - r_{x0} } \right)^{2\,} + \,\left( {y - r_{y0} } \right)^{2} } \right]}$} \!\mathord{\left/ {\vphantom {{\left[ {\left( {x - r_{x0} } \right)^{2\,} + \,\left( {y - r_{y0} } \right)^{2} } \right]} {\sigma^{2} }}}\right.\kern-0pt} \!\lower0.7ex\hbox{${\sigma^{2} }$}}}}$$
(7)

Feature selection by modified sea lion optimization

To improve the performance of classifier, feature selection is done which choose the relevant subset of features from the high dimensional dataset. The feature extracted by the above process occupies huge dimension which may reduce the effective performance of this proposed approach. In this method, a modified sea lion optimization algorithm [28] is applied for feature selection to avoid such issue. More relevant features are identified by this selection process and remove the unwanted features from the extracted features.

Modified sea lion optimization algorithm

One of the most intelligent animals is sea lion, which lives in tremendous colonies and it have a huge number of individuals. Within them there are numerous subgroups of their own hierarchy. It can move about these subgroups ordinarily in their lives. The navigations of sea lions rely upon their age, sexual orientation and capacity they have for the entire colony. The sea lions have three phases (i) using the whiskers the sea lion tracking and chasing the prey (ii) once it finds the prey the sea lion calling different individuals to joined their subgroup and surrounding the prey. And then the prey is attacked by sea lion.

Detecting and tracking phase

To recognize shape, position and size of features, the sea lions are utilized their whiskers. This encourages sea lion to detect the existing feature and to recognize their position when the whiskers course is on the other way of water waves. The position of the features is recognized by the sea lions and consider different individuals that will join its subgroup to locate the ideal features.

$$\mathop {Dist}\limits^{ \to } \, = \,\left| {\mathop {2B\,}\limits^{ \to \,} \,.\mathop {P(t)\,}\limits^{ \to \,} - \,\mathop {SL\,(t)\,}\limits^{ \to \,} \,} \right|$$
(8)

where, the distance between the target feature is indicated by \(\mathop {Dist}\limits^{ \to }\), the position vector of the target features and sea lion is represented as \(\mathop {SL\,(t)}\limits^{ \to }\) and \(\mathop {P\,(t)}\limits^{ \to }\), \(t\) is denoted as the current iteration and random vector is represented as \(\mathop B\limits^{ \to }\).

The sea lions move towards the target features to be nearest at the next iterations. The mathematical model of this behaviour is in below equation.

$$\mathop {SL\,(t + 1)}\limits^{ \to } \,\, = \,\,\mathop {P\,(t)}\limits^{ \to } \, - \,\mathop {Dist\,}\limits^{ \to } .\,\mathop C\limits^{ \to }$$
(9)

where, \(\mathop C\limits^{ \to }\) is decreased linearly from 2 to 0 and \(\mathop {(t + 1)}\limits^{{}}\) define the next iteration.

Finding best feature

Sea lions can detect the targets prey position and surround them. The hunting system is conducted by the leader (the best search agent), who finds prey and inform them concerning others. Typically target prey is measured the finest results for the current candidate.

$$\mathop {SL\,(t + 1)}\limits^{ \to } \,\, = \,\,\,\left| {\mathop {P\,(t)}\limits^{ \to } \, - \,\mathop {SL(t)}\limits^{ \to } } \right|\,\,\,.\,\,\cos \,(2\pi m)\, + \,\,\,\,\mathop {P\,(t)}\limits^{ \to }$$
(10)

For finding the best features, first we update the position by using galactic swarm optimization (GSO) algorithm [29]. GSO position is updated in sea lion optimization to get the optimal feature.

$$y^{(i)} \, \leftarrow \,y^{(i)} \, + \,v^{(i)}$$
(11)

where, \(y^{(i)}\) is represented as the position of GSO. \(\,\,\left| {\mathop {P\,(t)}\limits^{ \to } \, - \,\mathop {SL(t)}\limits^{ \to } } \right|\,\,\,\) denotes the distance among the finest optimum solution (target feature) and the search agent, random number is represented by \(m\) and \(\cos \,(2\pi m)\) is utilized to represent the features behaviour.

Recurrent fuzzy neural network classifier using social ski-driver for classification

In numerous fields, fuzzy neural networks have been used and RFNN is a notable fuzzy neural network. RFNN [30] has four layers such as input, membership, fuzzy and output layer. Let the input and output is \(o_{i}^{(k)}\) and \(u_{i}^{(k)}\) at the ith node in the layer \(k\) respectively. The following procedure is RFNN which is described in detail (Table 2).

Table 2 Hyper-parameters

Layer 1

First layer is called as input it has \(N\) number of nodes every layer is resembles with a parameter. Climate data input parameters like average humidity, evaporation per day, rain quantity, average of temperature and number of daylight hours.

$$O_{i}^{(1)} \, = \,\,u_{i}^{(1)} \, = \,x_{i} \,(t),\quad {\text{where}}\,\,i\,\, = \,\,1\, \div \,\,N$$
(12)

Layer 2

Membership layer is known as second layer. In fuzzy data, nodes in second layer will be changed to the crisp data using membership functions like Gauss function. In this layer, neural nodes is represented as \(N \times M\), where fuzzy rules is denoted by \(M\). Three parameters are in each and every nodes that is \(\theta_{ij}\), \(\sigma_{ij}\) and \(m_{ij}\).

$$O_{ij}^{(2)} \, = \,\exp \,\left[ { - \frac{{\left( {u_{ij}^{(2)} - m_{ij} } \right)^{2} }}{{\left( {\sigma_{ij} } \right)}}} \right],\quad where\;i\; = \;1\; \div \;N\;\;{\text{and}}\;\,j\; = \;1 \div M$$
(13)

where, \(m_{ij}\) is represented as the center of Gauss distribution function and \(\sigma_{ij}\) is represented as the variance of Gauss distribution function.

$$u_{ij}^{(2)} (t) = O_{i}^{(1)} \, + \,\theta_{ij} \,O_{ij}^{(2)} \,(t - 1),\quad {\text{where}}\,\,i = 1 \div N\;{\text{and}}\; = 1 \div M$$
(14)

where, \(O_{ij}^{{}}\) represents the weight of recurrent nodes. In this layer, the nodes from input has the factor \(O_{j}^{(2)\,} \,(t - 1)\,\). The residual data of the earlier learning step is denoted by this layer. This factor signifies the residual data of the earlier learning step. Therefore, in Eq. (13) after replacing \(u_{ij}^{(2)}\) by Eq. (12), we get Eq. (14).

$$O_{ij}^{(2)} \, = \,\,\exp \,\left[ { - \,\,\frac{{\left[ {x_{i} \,(t)\, + \,\theta_{ij} \,O_{ij}^{2} \,(t - 1)\, - \,m_{ij} } \right]^{2} }}{{\left( {\sigma_{ij} } \right)^{2} }}} \right]$$
(15)

Layer 3

Fuzzy rules is represented as the third layer. Fuzzy conclusion is given by connecting layer 3 and Layer 4. In this layer, each node resembles with an AND operation. Every AND operation expression is explained as below.

$$O_{j}^{(2)} \, = \,\,\prod\limits_{i = 1}^{N} {O_{ij}^{(2)} }$$
(16)
$$O_{j}^{(2)} \, = \,\prod\limits_{i = 1}^{N} {\exp \,\left[ { - \,\,\frac{{\left[ {x_{i} \,(t)\, + \,\theta_{ij} \,O_{ij}^{2} \,(t - 1)\, - \,m_{ij} } \right]^{2} }}{{\left( {\sigma_{ij} } \right)^{2} }}} \right]} \quad {\text{where}},\,j\, = \,1 \div \,M$$
(17)

Layer 4

Output layer is present in fourth layer which includes \(P\) nodes. \(P\) will be set to 1 in this model; this is the river runoff value. The fourth layer node is liable for changing fuzzy to crisp.

$$y_{k} \, = \,O_{k}^{(4)} \, = \,\sum\limits_{j = 1}^{M} {u_{jk}^{(4)} \,w_{jk} } \, = \,\sum\limits_{j = 1}^{M} {O_{jk}^{(3)} \,w_{jk} }$$
(18)
$$y_{k} \, = \,\sum\limits_{j = 1}^{M} {w_{jk} } \prod\limits_{i = 1}^{N} {\exp \,\left[ { - \,\frac{{\left[ {x_{i} \,(t)\, + \,\theta_{ij} \,O_{ij}^{2} \,(t - 1)\, - \,m_{ij} } \right]^{2} }}{{\left( {\sigma_{ij} } \right)^{2} }}} \right]} \quad {\text{where}},\,k\, = \,1 \div \,P$$
(19)

We use SSD [31] optimization algorithm to train RFNN after the defining process of RFNN. From various meta-heuristic algorithms, the behaviour of SSD was inspired. Its name tributes to the way that its stochastic exploration by one way or another resembles the ways that ski-drivers take downhill. Various parameters are in SSD, the description of these parameters is given below.

Position of the agent

To calculate the objective function, the position of the agents is utilized at that particular location, where dimension of search space is represented by n.

Previous best position

Using fitness function, the fitness value is designed for all agents. For each agent the fitness value is linked with its current position, and finally the best position is kept.

Mean global solution

Agents move to the global perspective, which speaks to the best three average solutions.

$$M_{i}^{t} \, = \,\frac{{X_{\alpha } \, + \,X_{\beta } \, + \,X_{\gamma } }}{3}$$
(20)

where, best three solutions are represented as \(X_{\alpha } \,,\,X_{\beta } \;{\text{and}}\;X_{\gamma }\)

Velocity of the agents

By adding the velocity \(V_{i}\), the agent positions are updated.

$$X_{i}^{t + 1} \, = \,X_{i}^{t} \, + \,V_{i}^{t}$$
(21)

where, \(V_{i}\) is the velocity of \(X_{i}\).

figure b

Results and discussion

The emotion recognition from group video is extensively applied in various applications. The whole work is done by Python and the performance is also carried out. The dataset is not available in publicly so we collected the video from YouTube. In this group video, the kids are introduced by their self and talking about their inventions. The dataset has size—6s video, source—YouTube, count—8 number of kids and statistics of classes—4 expressions. The accuracy, sensitivity, specificity, recall and precision is evaluated for four different emotions (Happy, sad, Fear and Angry) and the performance is compared with existing deep learning methods such as CNN [24], DNN + Bayesian [22], SVM [21] and CNN + VGG-16 [23]. The emotion recognition is shown in Fig. 5. The performance using GLCM, LESH and MLTP is shown in Fig. 6.

Fig. 5
figure 5

Flow chart of FFNN with SSD

Fig. 6
figure 6

Emotion recognition

Figure 7 shows the specificity performance with proposed and exsisting techniques. Here, in this graph the RFNN + SSD is denoted as orange color, CNN + VGG-16 is indicated as green color, SVM is denoted as ash color, DNN + Bayesian is denoted as brown color and CNN is denoted as pink color. Here, the specificity performance is taken as the average of 100% which includes the proposed and existing techniques. In that, the specificity performance of our proposed RFNN + SSD is very high i.e. 26.5% and the specificity performance of remaining existing deep learning methods like CNN + VGG-16 is 19.1%, SVM is 18.7%, DNN + Bayesian is 19.1% and the CNN is 16.6%. Thus, we know our proposed method achieved more specificity than others.

Fig. 7
figure 7

Specificity performance

Figure 8 shows the sensitivity performance with proposed and existing techniques. Here, the sensitivity performance is taken as the average of 100% which includes the proposed and existing techniques. In that, the sensitivity performance of our proposed RFNN + SSD is very high, i.e. 27.0% and the sensitivity performance of remaining existing deep learning methods like CNN + VGG-16 is 20.4%, SVM is 19.6%, DNN + Bayesian is 18.2% and the CNN is 14.8%. Thus, we know the performance of our proposed method achieved more sensitivity than others.

Fig. 8
figure 8

Sensitivity performance

Figure 9 shows the recall performance with proposed and existing techniques. Here, the recall performance is taken as the average of 100% which includes the proposed and existing techniques. In that, the recall performance of our proposed RFNN + SSD is very high i.e. 27.1% and the recall performance of remaining existing deep learning methods like CNN + VGG-16 is 20.4%, SVM is 19.6%, DNN + Bayesian is 18.2% and the CNN is 14.8%. Thus, we know the performance of our proposed method achieved more recall than others.

Fig. 9
figure 9

Recall performance

Figure 10 shows the precision performance with proposed and exsisting techniques. Here, the precision performance is taken as the average of 100% which includes the proposed and existing techniques. In that, the precision performance of our proposed RFNN + SSD is very high i.e. 25.8% and the precision performance of remaining existing deep learning methods like CNN + VGG-16 is 19.5%, SVM is 19.5%, DNN + Bayesian is 19.0% and the CNN is 16.1%. Thus, we know the performance of our proposed method achieved more precision than others.

Fig. 10
figure 10

Precision performance

Figure 11 shows the accuracy performance with proposed and existing methods. Here, the accuracy performance is taken as the average of 100% which includes the proposed and existing techniques. In that, the accuracy performance of our proposed RFNN + SSD is very high i.e. 27.1% and the accuracy performance of remaining existing deep learning methods like CNN + VGG-16 is 19.8%, SVM is 19.5%, DNN + Bayesian is 18.5% and the CNN is 15.1%. Thus, we know the performance of our proposed method achieved more accuracy than others.

Fig. 11
figure 11

Comparison of accuracy

Figure 12 shows the proposed method performance. Here, the accuracy of our proposed method is 99.16%, recall is 99.33%, specificity is 99%, precision is 99% and the sensitivity is 99.93%.

Fig. 12
figure 12

Performance of proposed methods

Figure 13 shows the performance of proposed RMSE, FNR and FPR. The proposed method contains less root mean square error (RMSE), false negative rate (FNR) and false positive rate (FPR). The value of proposed RMSE is 0.091%, FPR is 0.01% and the FNR is 0.0066%.

Fig. 13
figure 13

Proposed RMSE, FNR and FPR

Figure 14 shows the proposed performance of accuracy, recall, specificity, precision and sensitivity for low resolution images. The proposed accuracy value for low resolution image is 87.8%, recall value of low resolution image is 88%, specificity value for low resolution is 87.7%, precision for low resolution is 87.7% and sensitivity value for low resolution is 88%.

Fig. 14
figure 14

Low resolution for accuracy, recall, specificity, precision and sensitivity

Figure 15 shows the proposed performance of RMSE, FPR and FNR for low resolution images. The proposed performance of RMSE value for low resolution is 0.35%, the FPR value for low resolution image is 0.13% and the FNR value for low resolution image is 0.124%.

Fig. 15
figure 15

Low resolution for RMSE, FPR and FNR

Conclusion

In this proposed work, an emotion recognition from group video is taken as the major research topic. In different fields, group video based emotion recognition is gaining huge significance. In different manner, various emotions are shown by human, therefore it is difficult for machines to accurately recognize the human emotions. So here, a group based emotion recognition from video sequence with hybrid optimization based recurrent fuzzy neural network is proposed. In our work, emotion recognition from group of people is performed by utilizing the Neural Network. From the multi user video data, initially, original video frames are taken as input and pre-process it. From this pre-processed image, extract the facial features using MLTP, GLCM, and LESH and carried out as the features in the further processing steps. After extracting the features, certain features are selected using Modified Sea-lion optimization algorithm process. Finally, RFNN classifier using SSD, optimization algorithm is proposed for updating the weights in the RFNN. The whole work is done by Python platform and the performance of accuracy, sensitivity, specificity, recall and precision is evaluated with some existing techniques. The proposed method accuracy is 99.16%, recall is 99.33%, precision is 99%, sensitivity is 99.93% and specificity is 99% when compared with other deep learning techniques our proposed method attains good result. For low resolution image, the proposed method accuracy is 87.8%, recall is 88%, specificity is 87.7%, precision is 87.7% and sensitivity is 88%. Here, the data is taken from YouTube, when we compared with more data’s the complexity will be increase. So in future we will increase the number of data and reduce the complexity by using different machine learning techniques.