Introduction

Computer vision (CV) is an artificial intelligence-aided paradigm that is used to detect the digital world via cameras, videos, and deep-learning methods [1]. It classifies objects in an appropriate manner and responds to them automatically. CV focuses on the 3D modeling of objects, multiple-camera geometric analyses, cloud processing, and inferences based on the motion [2]. It acquires input from the machine and processes the output based on specific knowledge, namely, object labels and synchronization. It also includes developing particular technologies, such as image recognition, visual recognition, and facial recognition [3]. Mostly, CV is used to achieve high-level understanding of digital processing. The task of CV is to capture, examine, and recognize digital objects and extracts in higher dimensions [4]. This task allows it to provide the scientific or representative information needed for the transformation of images used in geometry, physics, statistics, and learning theory [5].

Human–machine interactions (HMIs) involve communicating and cooperating with the machine through a user interface. These interactions are carried out between the user and the machine to control the machine’s intuitive behaviors [6]. In recent years, specific sensors have been used to capture the normal, abnormal, and neural postures necessary for controlling the machine [7]. In HMIs, CV is used to acquire high-level photos and videos used in the understanding of these postures/images. CV is related to HMI in that they both detect and monitor objects as well as control them by determining the conditions under which they operate [8]. One of CV’s upgraded applications is hand-gesture recognition, which monitors the postures used in HMIs and interacts with them. The goal of this application is to obtain vigorous non-specific vision [9]. It is used in automatic parking, controlling music with a gesture, eye tracking with multiple touches, and so forth. CV and HMI are commonly used together in virtual reality applications. In comparison to other baseline technologies, CV combined with HMI produces the smallest errors [10].

The data processing in CV is done by acquiring an image, then extracting the information that needs to be processed. CV obtains the input images using a high-level approach; based on this approach, classifications are attained that render the appropriate results [11]. Thus, recognition methods are developed to detect the images. The obtained images interact with the machine through HMI applications, such as assistance devices relying on voice and hearing [12], developed for specific platforms. Many prediction-based methodologies and machine-learning algorithms have been developed for CV detection that are related to HMI [13, 14]. The current work addresses the errors that occur when combining structured and unstructured data. It uses the tree method for classification, which allows it to attain maximum gain, and regression is used to reduce the error in combining the structured and unstructured data, maximizing the precision. The main contributions of this paper are as follows:

  • We maximize the structured and unstructured data classification accuracy by applying the classification and regression approach.

  • We reduce the error rate using two-level visual information processing (2LVIP) techniques, which help to minimize the misclassification rate.

  • Furthermore, we obtain the maximum information gain value for both the structured and unstructured data.

  • Finally, we improve the precision by classifying the structured and unstructured data and reduce the error by combining the classification and regression methods.

The rest of the paper is arranged as follows. “Related works” describes the various research options in terms of vision-based information processing, “Proposed 2LVIP method” explains the 2LVIP working process, “Results and discussion” evaluates the efficiency of the 2LVIP system, and conclusions are made in “Conclusion”.

Related works

Monocular robot vision (MRV) was proposed by Chan and Riek [15] for unseen salient object detection done in parallel with discovery prediction. Unsupervised foraging of objects (UFO) is the fastest and most accurate method for notable object discovery. It is done via the real-world perceptions of robots. The main intention is to improve autonomy and resolve robotic challenges while engaging in a salient object discovery process. The embedded computer-vision system in traffic surveillance was introduced by Mhalla et al. [16] for multi-object detection. This method, which is used for detecting traffic objects in traffic scenarios, consists of a robust detector that makes use of a generic deep detector and enhances detection accuracy.

Wang et al. [17] developed a scale-aware rotating-object detection system at low-level high resolution for obtaining high-level semantic information with aerial imagery. Intersection-over-union (IoU) loss coupled with scale diversity detects orientation. The proposed method is used to improve the accuracy of the rotating bounding box. Kulik et al. [18] addressed CV for intelligent robots by proposing a convolutional neural network for object detection that detects flags indicating unsatisfactory results for different objects. Training and the testing of objects is maintained throughout.

Shin et al. [19] equipped unmanned surface vehicles with object detection and tracking abilities to improve accuracy. The proposed system contributes to an extensive baseline stereo vision system designed to enhance sea surface estimation. It is applicable for long-range object detection, and the semantic segmentation required for detecting objects is done via the oblique convolution designed by Lin et al. [20]. The artificial intelligence developed for the CV uses pixel classification. The hourglass network analyzes the local extremes.

Maggipinto et al. [21] modeled two-dimensional data for CV virtual metrology (VM) using deep learning. The model is used for automatic feature extraction to improve the accuracy and scalability of the VM. The modeling is done using both spatial and time evolutions on real industrial data, including data semiconductor manufacturing. Luo et al. [22] introduced a vision-based detection system for a dynamic workspace involving workers on foot. Multiple detections are used for object tracking and action recognition, and the system determines two types of action data, such as classes and locations. A density-based spatial clustering algorithm is then used to analyze the dynamic workspace.

Jiang et al. [23] proposed fusing spatiotemporal data for hydrological modeling with vision-based data. Three steps are used in this model. The first step involves the fusion of multi-source spatiotemporal for the incorporation of big data. The second step is associated with short- and long-term forecasting, whereas the third step models the streamflow. A multi-object detection (MOD) method is introduced for autonomous vehicle applications that fuses three-dimensional light detection and ranging (LIDAR) and camera data. Zhao et al. [24] provided solutions for recognizing objects by identifying regions of interest (ROI) in the initial processing stage. Later, a convolutional neural network (CNN) is adopted for the recognition of objects. Sliding windows are used for the candidate object region detection of real-time autonomous vehicles. The introduced system maximizes the object detection region and minimizes the misclassification error rate.

Liu et al. [25] implemented a reference frame Kanade–Lucas–Tomasi (RF-KLT) algorithm for extracting the features in fixed regions. The dimensions of the features are reduced to detect the class boundaries. This work was done in a real-time environment, allowing the efficiency of the system to be evaluated with the help of an augmented and actual video dataset. The system was able to classify anomalies successfully in a robust and cost-effective manner. Fang et al. [26] proposed integrating CV with an ontology that can be used to identify the hazards on construction sites using a knowledge graph. Shu et al. [27] introduced a human–computer interaction mode through the interactive design of intelligent machine vision. Their objective was to improve the accuracy of the functioning algorithm, and the point-and-click results are based on Fitts’ Law.

In this work, two-level visual information processing is considered for performing semantic object detection. The efficiency of the system is evaluated against the methods used in [15, 24], and [25] because these methods work perfectly while analyzing objects. In other words, they are effective in detecting objects and regions in a robust and cost-effective manner. In addition, these methods can be used to evaluate real-time datasets and improve the object and region recognition process.

Proposed 2LVIP method

CV is used for learning the instances of semantic objects in an automatic detection manner. This paper’s objective is to improve the precision of classifying structured and unstructured data together, thus reducing the error introduced by combining these classification methods. It uses regression methods to do so; that is, the proposed method uses two-level visual information processing (2LVIP) to obtain the maximum gain from the inputs. Figure 1 depicts the proposed model.

Fig. 1
figure 1

HMI using 2LVIP

Figure 1 clearly depicts the overall architecture of the HMI system with 2LVIP. The imaging devices are used to gather information from the environment. The collected information is then processed by the 2LVIP. The regression and classification techniques are incorporated into the 2LVIP to maximize unstructured and structured data processing. The proposed processing method identifies the gain-related feature information in the data in the first level and optimizes it to maximize the gain. In the second level, the errors in the classifications are determined, and they are removed via the regression process, thus stabilizing the precision to meet the HMI application’s demands. The following equation represents the structured and unstructured images used as the input for the CV system:

$${l}_{0}=\sum_{{t}_{a}}^{{d}_{t}}\left(1+\frac{{i}_{0}}{{\mathrm{ni}}_{0}}\right)*{c}_{0}+\sqrt{\left(\frac{{s}_{r}+{u}_{r}}{{\mathrm{ni}}_{0}}\right)}.$$
(1)

Equation (1) indicates the analysis \({l}_{0}\) of structured \({s}_{r}\) and unstructured \({u}_{r}\) images, where the structured image includes the size of the image and presentation, whereas the unstructured image includes variations in size, frame, and patterns. The structured image remains the same for the single input image when a number of images \(\left\{{i}_{0}^{1},{i}_{0}^{2},\ldots {\mathrm{ni}}_{0}\right\} \text{are }\, \text{used}.\) In this case, \({\mathrm{ni}}_{0}\) represents the number of images that are captured at the appropriate time \({t}_{a}\). The time the image is captured is denoted as \({d}_{t} ({c}_{0})\). After the classification of the structured and unstructured data is done using the tree model. Equation (2a) below indicates the grouping of the trees:

$$\theta =\left.\begin{array}{c}{i}_{0}+\prod\limits_{{c}_{0}}^{n{i}_{0}}\left(1+\frac{{i}_{0}}{n{i}_{0}}\right)*{s}_{r} \forall \, \mathrm{structured} \,\mathrm{data}\\ \left({c}_{0}+n{i}_{0}\right)*\sqrt{\frac{{i}_{0}}{\sum_{{t}_{a}}{c}_{0}}}+{u}_{r} \forall \, \text{unstructured}\, \text{data}\end{array}\right\}.$$
(2a)

Object detection occurs for both structured and unstructured data, where \(1+\frac{{i}_{0}}{{\mathrm{ni}}_{0}}*{s}_{r}\) represents the structured image data. The analysis of the unstructured data is denoted as \(\sqrt{\frac{{i}_{0}}{\sum_{{t}_{a}}{l}_{0}}}+{u}_{r}\). The combination of these two parts contains the necessary information. Equation (2b) is used to obtain useful information:

$$\beta =\left[{l}_{0}\left({i}_{0}\right)+\prod\limits_{f^{\prime}}^{{c}_{0}}{\mathrm{ni}}_{0}+\frac{{\mathrm{ni}}_{0}}{\sum {s}_{r}}*\left({d}_{t}-{t}_{a}\right)\right]+\left(1+\frac{{c}_{0}+({u}_{r}-{s}_{r})}{{\mathrm{ni}}_{0}}\right).$$
(2b)

The extraction of useful information \(\beta \) from the data is the initial step for classification here; the analysis is done by evaluating Eq. (2b). The data processing is illustrated in Fig. 2.

Fig. 2
figure 2

Input data processing

In Fig. 2, the processing is carried out for the structured and unstructured data via \(\frac{{\mathrm{ni}}_{0}}{\sum {s}_{r}}*\left({d}_{t}-{t}_{a}\right)\). Thus, it considers the two types of data with based on the classification based on the mathematical computation \(1+\frac{{c}_{0}+({u}_{r}-{s}_{r})}{{\mathrm{ni}}_{0}}\) and examines the data that are necessary for automatic detection in a real-time environment.

Classification via the discrete finite value

The discrete finite value is used to derive the classifications for structured and unstructured data. It is based on data that are acquired at a particular time and indicates two sets, such as the finite \({\partial }_{0}\, \text{ and}\, \text{infinite}\, \partial ^{\prime}\). Thus, by combining Eqs. (2a) and (2b), Eq. (3) below is derived for identifying the data in \({f}^{^{\prime}}\):

$$\alpha \left(\beta \right)=\left\{\begin{array}{l}\left(1+\frac{{i}_{0}}{{\mathrm{ni}}_{0}}\right)*{s}_{r}+{i}_{0}\\ \text{such}\, \text{ that}\, {s}_{r}\in {f}^{^{\prime}}+{c}_{0}\\ {l}_{0}\left({i}_{0}\right)+\left(1+\frac{{c}_{0}+\left({u}_{r}-{s}_{r}\right)}{{\mathrm{ni}}_{0}}\right)*\left({d}_{t}-{t}_{a}\right)\\ \text{such }\, \text{that }{u}_{r}\in \left(1+\frac{{l}_{a}}{{\mathrm{ni}}_{0}}\right)\end{array}.\right.$$
(3)

The useful information is extracted for the classification \(\alpha \left(\beta \right)\) is formulated above. The structured data \({s}_{r}\in {f}^{^{\prime}}+{c}_{0}\) is identified when the image is captured. The unstructured \({u}_{r}\in \left(1+\frac{{l}_{a}}{{\mathrm{ni}}_{0}}\right)\) is used in the analysis of a number of input images. The information is extracted in this step, and a discrete set of values is identified. Equation (4a) below is then used to determine the classification process for this discrete set of values:

$$\begin{aligned} \alpha \left({l}_{0}\right) & = \prod\limits_{{t}_{a}}^{{d}_{t}}{l}_{0}+\beta *{(\partial }_{0}+\partial ^{\prime})\left[\left(1+\frac{{c}_{0}+\left({u}_{r}-{s}_{r}\right)}{{\mathrm{ni}}_{0}}\right) \right. \\ &\quad \left. +\left(\sqrt{\frac{{i}_{0}}{\sum_{{t}_{a}}{l}_{0}}}+{\mathrm{ni}}_{0}\right)*{(c}_{0}+{i}_{0})+\left(\sqrt{\left(\frac{{s}_{r}+{u}_{r}}{{\mathrm{ni}}_{0}}\right)}+\left({s}_{r}+{u}_{r}\right)\right)\right] \\ &\quad +\left(1+\frac{{l}_{a}}{{\mathrm{ni}}_{0}}\right)*({t}_{a}-{d}_{t}).\end{aligned} $$
(4a)

Datasets are either finite or infinite based on this classification scheme. The data belonging to finite sets are distinguished as \({\partial }_{0}\left({i}_{0}+{l}_{0}\right);\) these are the data that are acquired at a particular time. If data are not gathered on time, then it is considered to be unstructured data. Equation (4b) is then used to represents the finite and infinite classifications:

$$\alpha =\left\{\begin{array}{l}\left.\begin{array}{l}\sum\limits_{{l}_{0}}^{{c}_{0}}\left({l}_{0}+\frac{{f}^{\prime}}{{\mathrm{ni}}_{0}}\right)*\frac{{e}^{\prime}}{{s}_{r}}\\ \frac{{i}_{0}}{\sum_{{t}_{a}}{l}_{0}}*\prod\limits_{{t}_{a}}{(c}_{0}+e^{\prime})\end{array}\right\}, {\partial }_{0}\\ \left.\begin{array}{l}\prod\limits_{{d}_{t}}^{{c}_{0}}\beta +{l}_{0}*\frac{{u}_{r}}{{\mathrm{ni}}_{0}}\\ \left(1+\frac{{l}_{a}}{{\mathrm{ni}}_{0}}\right)+{(c}_{0}+{i}_{0})+\prod\limits_{{d}_{t}}^{{i}_{0}}{u}_{r}+e^{\prime}\end{array}\right\},\partial ^{\prime}\end{array}\right..$$
(4b)

For discrete values \(e^{\prime}\), the classification is carried out by observing if the data are processed at fixed times with the structure \(\frac{{e}^{\prime}}{{s}_{r}}+\left({l}_{0}*{f}^{\prime}\right)+({d}_{t}-{t}_{a})\). The unstructured data that are not processed at the specific time are represented by \(\left({l}_{0}*\frac{{u}_{r}}{{\mathrm{ni}}_{0}}\right)+\left({u}_{r}+{e}^{^{\prime}}\right)-{t}_{a}\). The classification is then carried out and attains maximum gain. The maximum gain is obtained by combining the structured data \({s}_{r}\, \text{and}\, \text{ unstructured }\, \text{data}\, {u}_{r}\). The error that occurs during processing is \({l}_{0}+({s}_{r}+{u}_{r})\). The error is due to the misclassification of discrete values (\({\alpha }_{0})\), which is computed using Eq. (4c):

$$ \begin{aligned} {\alpha }_{0} & =\left[\left[\left(\left({l}_{0}+\frac{{f}^{\prime}}{{\mathrm{ni}}_{0}}\right)*\left({e}^{\prime}+\beta \right)+\frac{{i}_{0}}{\sum_{{t}_{a}}{l}_{0}}\right)+ {\raise0.7ex\hbox{${d_{t} }$} \!\mathord{\left/ {\vphantom {{d_{t} } {f^{\prime}}}}\right.\kern-\nulldelimiterspace} \!\lower0.7ex\hbox{${f^{\prime}}$}} \right]*{s}_{r} \right. \\ &\quad \left. +\left[\left({t}_{a}*{c}_{0}\right)*\left({l}_{0}*\frac{f^{\prime}}{{\mathrm{ni}}_{0}}\right)+\left({l}_{0}+\beta \right)\right]*{u}_{r}\right]-\left({d}_{t}-{f}^{\prime}\right).\end{aligned}$$
(4c)

A misclassification leads to a finite value \({e}^{^{\prime}}\left({\partial }_{0}\right)\, \text{ that}\, \text{ is }\, \text{ not }\, \text{ discrete}\), as observed in Eq. (4c). In these cases, \(\left({l}_{0}+\frac{{f}^{\prime}}{{\mathrm{ni}}_{0}}\right)*\left({e}^{\prime}+\beta \right)\) represents the discrete value that extracts useful information, although the identification is not made on time. The classification process is illustrated in Fig. 3.

Fig. 3
figure 3

Classification process

The computation of \({(l}_{0}+\beta )*{u}_{r}\) results in an infinite set of values. Following this process, the accurate classification of the finite set is examined, and the error is reduced by deriving the regression algorithm.

Minimizing the errors with regression

The regression method has been used to predict the structured and unstructured data to reduce \(\text{the}\, \text{ error }\, \text{ caused }\, \text{ at}\, \text{ the}\, \text{ time}\, \text{ of}\, \text{ classification} \,\partial ^{\prime}\). This second-level extracts better precision from the HMI application. The process obtained in the first level extracts the maximum gain by evaluating both \({s}_{r }\, \text{and} \, {u}_{r}.\) During this process, some errors occur. A prediction is made by observing the training set that is processed in the preceding process. Equation (5a) below is used to evaluate the regression through prediction:

$${q}_{0}=\left\{\begin{array}{l}\sum\limits_{{l}_{0}}^{{c}_{0}}\frac{f^{\prime}}{{\mathrm{ni}}_{0}}+\left(\beta +\alpha \right)*{t}_{a}\\ \left(\alpha +\sqrt{\frac{{l}_{0}}{\prod_{{c}_{0}}{i}_{0}}}\right)*\left({s}_{r}+{u}_{r}\right)+\sum\limits_{{t}_{a}}^{{l}_{0}}\left({f}^{\prime}+\beta \right).\end{array}\right.$$
(5a)

The regression allows the prediction of \({\partial }_{0} \, \text{and}\, \partial ^{\prime}\). The dependent value is used to find the prediction \({q}_{0}\) used in \(\left[\frac{f^{\prime}}{{\mathrm{ni}}_{0}}+\left(\beta +\alpha \right)\right]\) to identify the unstructured data. Here, \({\sum}_{{t}_{a}}^{{l}_{0}}\left({f}^{\prime}+\beta \right)\) represents the useful information that is extracted for further processing; the independent value is derived via Eq. (5b) below:

$${p}^{^{\prime}}=\prod\limits_{\alpha }^{{l}_{0}}\left(1+\frac{{s}_{r}+{u}_{r}}{\text{n{i}}_{0}}\right)*\sum_{\beta =1}^{{\alpha }_{0}}{(l}_{0}+f^{\prime})$$
(5b)

Here, \(1+\frac{{s}_{r}+{u}_{r}}{{\mathrm{ni}}_{0}}\) denotes the structured and unstructured data that are combined to attain the maximum gain. The prediction method is formulated in Eq. (6), as follows:

$$ \begin{aligned} \gamma & ={o}^{\prime}\left({g}_{0}\right)-{t}_{a}*\left\{\left[\prod\limits_{{q}_{0}}^{{p}^{\prime}}\left(\frac{\beta \left({i}_{0}\right)}{\alpha }*{c}_{0}\right)+\sum\limits_{{s}_{r}}^{{u}_{r}}\left({t}_{a}-\beta \right)\right] \right. \\ & \quad \left. +\left[{(\partial }_{0}+\partial ^{\prime})+\left( {\raise0.7ex\hbox{${l_{0} }$} \!\mathord{\left/ {\vphantom {{l_{0} } {f^{\prime}}}}\right.\kern-\nulldelimiterspace} \!\lower0.7ex\hbox{${f^{\prime}}$}} \right)-{(d}_{t}-e^{\prime})+\prod\limits_{\beta }^{{e}^{\prime}}{c}_{0}+{i}_{0}\right]\right\}+\left({g}_{0}-{h}_{0}\right).\end{aligned}$$
(6)

Here, \(\frac{\beta ({i}_{0})}{\alpha }*{c}_{0}\) is used to gain information for the input image. Figure 4 illustrates the regression process.

Fig. 4
figure 4

Illustration of the regression process

It determines the \(\left({t}_{a}-\beta \right)\) targeted for extraction from the obtained gain at a particular time. The error \(o^{\prime}\) is used to examine the foregone and forthcoming data for the HMI applications \({g}_{0}\, \mathrm{and}\, {h}_{0}\). In this manner, the error of the foregone data is compared with that for the forthcoming data to provide an optimal result. The objective of the second level is to reduce the errors made in information processing. Thus, the training set is used to compare and improve processing further, see Eq. (7).

$$\varnothing =\left\{\begin{array}{l}1 \, \text{if} \, \sum\limits_{{q}_{0}}^{p^{\prime}}\Vert {(l}_{0}+f^{\prime})*{i}_{0}\Vert +\left(\frac{\beta +e^{\prime}}{{i}_{0}}\right)+({g}_{0}-{h}_{0})\\ 0, \text{otherwise}\end{array}\right..$$
(7)

The training data \(\varnothing \) are used to formulate the ‘if and otherwise’ conditions for \(\Vert {(l}_{0}+f^{\prime})*{i}_{0}\Vert \), thus \(\frac{\beta +e^{\prime}}{{i}_{0}}\) is used for identifying the discrete data. The foregone and forthcoming data are used to evaluate this prediction-based method. The foregone data is again trained and used for the forthcoming data based on the ‘if’ condition. The training calculation in Eq. (7) is used to make the comparison needed for the HMI application. The cost function calculated for the regression method minimizes the error between the predicted and actual values used for processing the data. Equation (8) below is used to obtain the cost function:

$$\Delta =\frac{1}{j^{\prime}}\prod\limits_{\alpha }^{j^{\prime}}(\gamma {-{h}_{0})}^{2}+\left(\beta +\frac{{p}^{\prime}+{q}_{0}}{{i}_{0}}\right).$$
(8)

The cost function \(\Delta \) is used to attain the data points \({j}^{\prime}\) used to find the prediction value, then forthcoming data is subtracted, along with the prediction value. The \(\left(\beta +\frac{{p}^{\prime}+{q}_{0}}{{i}_{0}}\right)\) is processed to obtain the actual value. The cost function is used for obtaining better results for the classification and regression method. Next, the prediction-based regression method is used to reduce the error with Eq. (9), as follows:

$$\rho =\left\{\begin{array}{l}\left(\Delta +\frac{\beta +e^{\prime}}{{i}_{0}}\right)*\left({g}_{0}-{h}_{0}\right)+\left(\alpha +\beta \right)=1\\ \prod\limits_{\alpha }^{{l}_{0}}\left({f}^{\prime}+{i}_{0}\right)*\left(\frac{\beta +\gamma }{\varnothing }+\Delta \right)\left({g}_{0}-{h}_{0}\right)\ne 1\end{array}\right..$$
(9)

The reduction in the error \(\rho \) made in Eq. (9) is used to evaluate the error. The cost function \(\Delta +\frac{\beta +e^{\prime}}{{i}_{0}}\) is used in the analysis of the regression method. The training set is used to determine the data from the maximum gain \(\left(\frac{\beta +\gamma }{\varnothing }+\Delta \right)\left({g}_{0}-{h}_{0}\right).\) It derives the lesser error. Afterwards, a final regression method is used to reduce the error from the classification of the structured and unstructured data, see Eq. (10):

$$\delta \left({i}_{0}\right)=\left\{\begin{array}{c}\left[\left(\beta +\frac{{p}^{\prime}+{q}_{0}}{{i}_{0}}\right)+{(l}_{0}+f^{\prime})*\prod\limits_{\beta }^{e^{\prime}}\left(\alpha +\gamma \right)\right]-{d}_{t}<o^{\prime}\\ \sum\limits_{{l}_{0}}^{\alpha }\left(\beta +\gamma \right)+{(l}_{0}+f^{\prime})+\left(1+\frac{{l}_{a}}{{\mathrm{ni}}_{0}}\right)*({t}_{a}-{d}_{t})>o^{\prime}\end{array}.\right.$$
(10)

In this regression, \(\left(\alpha +\gamma \right)-{d}_{t}\) has more errors at the time of computing. Figure 5 illustrates this regression-based detection method.

Fig. 5
figure 5

Regression-based detection process

The second condition \((\beta +\gamma )*({t}_{a}-{d}_{t})\) has a lesser error. In this equation, the first condition does not satisfy the objective, whereas the second condition satisfies the proposed method (2LVIP). If an object is not classified correctly, the derivative \(\frac{{f}^{^{\prime}}+\beta }{{\mathrm{ni}}_{0}}*{l}_{0}\) is applied. Thus, the process is updated for better classification. As a result, the precision is enhanced, and error is reduced in the HMI application through the evaluation of \(\left(\alpha +\gamma \right)*\delta \left({i}_{0}\right)\).

Results and discussion

This section discusses the performance evaluation of the proposed method that is done through a comparative study of different metrics. In this assessment, the metrics considered are precision, gain ratio, error, and processing time. The image input from [28] is used in this analysis. It consists of 102 sets of images, along with their annotations. A region’s image with 45 training data points is verified in the analysis, and the results are discussed. The size of the training dataset is 149 mb, and the annotations fill nearly 7 mb. The number of classification and regression instances considered are 20 and 40, respectively. The RF-KLT [25], MOD-3D LIDAR [24], and UFO [15] methods are used in the comparison.

Precision analysis

In Fig. 6a and b, the precisions of the classification instances and information gain ratios are compared. The precision for the proposed 2LVIP is higher due to the evaluation of \(\frac{{i}_{0}}{\sum_{{t}_{a}}{l}_{0}}\), i.e., the data obtained from the classification tree. The analysis of the unstructured data \(\left(1+\frac{{i}_{0}}{{\mathrm{ni}}_{0}}\right)*{s}_{r}+{u}_{r}\) determines the resultant data \(\left({c}_{0}+{\mathrm{ni}}_{0}\right)\). The analysis proceeds by introducing the regression methods to acquire the number of classification instances based on time \({\partial }_{0}+\frac{{s}_{r}+{u}_{r}}{{i}_{0}}\). The data so attained have higher classifications; thus, the gain also increases. The gain is determined by deriving \(\left(\alpha +\frac{f^{\prime}}{{i}_{0}}\right)+\frac{\beta }{e^{\prime}},\) which represents an actual stage in the processing. The data processing in this 2LVIP method is analyzed by evaluating \(\varnothing +{i}_{o}*({g}_{0}+{h}_{0})\). The foregone and forthcoming data are used in the analysis of the precision. By comparing the classification of the proposed method with those of the other three methods, it can be seen that a smaller precision value is obtained when the information gain decreases. Thus, if the classification is improved, then the gain also increases. Thus, the error reduces \(\left({i}_{0}+\frac{\alpha +f^{\prime}}{\beta }\right)*{t}_{a}-{d}_{t}\) and vice versa. In Fig. 6, a higher precision is obtained by computing \({t}_{a}+{f}^{\prime}+\beta \).

Fig. 6
figure 6

a Precision versus classification instances. b Precision versus information gain %

Information gain analysis

The gain for the proposed method is shown in Fig. 7, which indicates that as the classification increases, the gain increases. When comparing the 2LVIP method with the other three methods, namely, RF-KLT, MOD-3D LIDAR, UFO, the 2LVIP method has the highest gain. The gain performance is found through \(1+\frac{{c}_{0}+({u}_{r}-{s}_{r})}{{\mathrm{ni}}_{0}},\) which acquires the image and classifies the data. Improvements in the gain can be analyzed by determining \({s}_{r}\in {f}^{^{\prime}}+{c}_{0}\, \mathrm{and}\, {u}_{r}\in \left(1+\frac{{l}_{a}}{{\mathrm{ni}}_{0}}\right)\). The maximum gain is obtained when the classification of the structured and unstructured data is done correctly. The classification methods are derived in Eq. (4a), where \(\beta *{(\partial }_{0}+\partial ^{\prime}) \, \text{represents} \, \text{ the} \, \text{ extraction} \, \text{ of} \, \text{ the} \, \text{ useful } \, \text{information}\). The data belonging to finite sets are distinguished as \({\partial }_{0}({i}_{0}+{l}_{0})\). If the data are not collected promptly, then they are represented as unstructured data, namely, by \(\left({l}_{0}+\frac{{f}^{\prime}}{{\mathrm{ni}}_{0}}\right)*\frac{{e}^{\prime}}{{s}_{r}}\). These data are not processed at the specific time \(\left({l}_{0}*\frac{{u}_{r}}{{\mathrm{ni}}_{0}}\right)+\left({u}_{r}+{e}^{\prime}\right)-{t}_{a}\).

Fig. 7
figure 7

Gain % versus classification instances

Error analysis

In the 2LVIP method, the error is reduced by increasing the number of correct classifications at the specific time \(\left({l}_{0}*\frac{{u}_{r}}{{\mathrm{ni}}_{0}}\right)+\left({u}_{r}+{e}^{^{\prime}}\right)-{t}_{a}\). Thus, these classifications signify the useful information obtained and also the error that occurs during processing, namely, \({l}_{0}+({s}_{r}+{u}_{r})\). The error is due to the misclassification of discrete values. The expression \(1+\frac{{s}_{r}+{u}_{r}}{{\mathrm{ni}}_{0}}\) denotes the structured and unstructured data that are combined to attain the maximum gain. Next, the prediction \(\gamma +\beta \left({i}_{0}\right)*\alpha \) is determined to reduce the error in the forthcoming data. The analysis is done by computing the process at a specified time. Thus, the prediction is used to obtain errorless unstructured and structured data. Equation (9) is used to reduce the error, and \(\left(\Delta +\frac{\beta +e^{\prime}}{{i}_{0}}\right)*{f}^{^{\prime}}+{i}_{0}\) is derived. Thus, information gain obtained from the classification increases, so that the error is reduced. The error is smaller for the 2LVIP method than the errors for the existing three methods, as seen by evaluating the \({o}^{^{\prime}}+{f}^{^{\prime}}\left({i}_{0}\right)\) (refer to Fig. 8a, b).

Fig. 8
figure 8

a Error versus the classification instances. b Error versus information gain %

Analysis of the processing time

Although there is a high amount of data for classification when using the 2LVIP method, it has the smallest processing time, as shown in Fig. 9a. As the number of classifications increases, the analysis necessary increases. It is found by determining \(\left({t}_{a}*{c}_{0}\right)*\left({l}_{0}*\frac{{f}^{^{\prime}}}{{\mathrm{ni}}_{0}}\right),\) which evaluates the specific time needed for processing. Using Eq. (4a), the appropriate classification is carried out for the acquired data. If \(\left({l}_{0}+\frac{{f}^{^{\prime}}}{{\mathrm{ni}}_{0}}\right)*\left({e}^{^{\prime}}+\beta \right)\) represents the discrete value that extracts useful information and identification is not made on time, it must be improved. The data obtained from the classification reduces the error at the time of classification \({\partial }_{0},\, \text{thus} \, \text{ providing}\) a better processing time for the HMI application. The classification analysis uses \(\alpha \left({l}_{0}\right)*\left({t}_{a}-{d}_{t}\right),\) which extracts the data at the appropriate time. The data belonging to finite sets are distinguished as \({\partial }_{0}({i}_{0}+{l}_{0})\). If the information gain increases, then the processing time for the proposed method, formulated with \(\gamma +\beta \left({l}_{0}\right),\, \text{holds} \, \text{constant}\) (refer to Fig. 9a, b). The results of the above comparison are tabulated in Tables 1 and 2, respectively.

Fig. 9
figure 9

a Processing time versus classification instances. b Processing time versus information gain %

Table 1 Metrics for the different classification instances
Table 2 Metrics for the different gain %

The cumulative outcome of the proposed method shows that it improves precision and gain% by 3.61% and 9.42%, respectively. It also reduces the error and processing time by 6.51% and 21.75%, respectively.

In terms of the information gain, the proposed method achieves 4.54% higher precision, 7.65% less error, and 26.48% less processing time.

Analysis of the instances

In Fig. 10, the regression instances for the classified data increase, so the error decreases through the derivation of \(\frac{\beta ({i}_{0})}{\alpha }*{c}_{0}\). It determines \(\left({t}_{a}-\beta \right)\) for extraction at a particular time from the obtained gain. The error \(o^{\prime}\) is used to examine the foregone and forthcoming data for the HMI applications \({g}_{0}\, \mathrm{and} \,{h}_{0}\). Here, \(\Vert {(l}_{0}+f^{\prime})*{i}_{0}\Vert \) is used to identify the analysis of the input data. Thus, the \(\left(\beta +\frac{{p}^{^{\prime}}+{q}_{0}}{{i}_{0}}\right)\) are computed for better precision. The maximum gain \(\left(\frac{\beta +\gamma }{\varnothing }+\Delta \right)\left({g}_{0}-{h}_{0}\right)\) is used to reduce the error.

Fig. 10
figure 10

Precisions of the different regression and classification instances

The maximum gain is increased by finding \(\left( {\raise0.7ex\hbox{${l_{0} }$} \!\mathord{\left/ {\vphantom {{l_{0} } {f^{\prime}}}}\right.\kern-\nulldelimiterspace} \!\lower0.7ex\hbox{${f^{\prime}+ \beta}$}} \right)-{(d}_{t}-e^{\prime}),\) which is used to decrease the error. In Fig. 11, if the number of classifications the proposed method makes increases, then its error is reduced. For the 2LVIP method, the \(\left({g}_{0}-{h}_{0}\right)+\left(\alpha +\beta \right)\) obtained with the final regression method is used to reduce the error in the classification of structured and unstructured data. The reduction in the error \(\rho \) is found with Eq. (9), then used to evaluate the error.

Fig. 11
figure 11

Errors for the different regression and classification instances

Given the number of instances, the regression produces more gain than the classification; hence, the error is reduced. The classification of the structured and unstructured data is evaluated with the regression equation \(\delta \left({i}_{0}\right)\). Equation (10) is used to reduce the error and determines the regression process based on the prediction \(\gamma +\frac{{f}^{^{\prime}}+e^{\prime}}{{i}_{0}}\) (refer to Fig. 12).

Fig. 12
figure 12

Gain % for the different classification and regression instances

Conclusion

In this paper, a two-level visual information processing method is discussed for improving the precision of human–machine interaction systems. The input obtained from the imaging devices is classified into structured and unstructured data based on the available information. The maximum information gain is then extracted, and regression is used to mitigate the errors in the information gained. The regression process uses the training set data to reduce the error and processing time via recursive discrete and finite value estimations. The regression analysis is recursively handled using predictive cost estimation. Training is performed to reduce the processing time used in extracting useful information from the visual input. By attuning the classification and regression processes, the proposed method is found to maximize the precision and information gain in detecting the objects and reduce the error and processing time. In the future, an optimization-based regression approach will be applied to classify the objects.