Introduction

In the area of non-destructive testing and evaluation (NDT&E), infrared thermography (IRT) has evolved rapidly during the past decades thanks to the advances in infrared (IR) cameras and lenses. Among the industrially utilized IRT techniques, vibrothermography is a special application which usually relies on the friction inside the inspected structure as the heat source. In a vibrothermographic test, the inspected structure is excited using mechanical forces or acoustic/ultrasonic waves at a single frequency—or multiple frequencies—which induce structural vibration. During the vibration, friction is often generated within the structure due to relative motion between surfaces and components. Because of the friction-based heat generation mechanism, vibrothermography is known for its effectiveness in detecting cracks and delamination [1]. Another operational advantage of vibrothermography is that as heat is generated within the structure, the travel distance of the thermal wave is shortened compared to the case in conventional active IRT tests where heat flows are applied externally to the material surface.

Compared to other commonly deployed NDT&E techniques, IRT approaches, including vibrothermography, have distinctive advantages in operating conditions due to the capability to perform rapid non-contact scans of large areas with short measurement time. The results from IRT tests are usually presented in the form of high-definition two-dimensional images or videos generated from the temperature data acquired by IR cameras. These high-definition images or videos are able to provide clear and robust demonstrations of the condition of the inspected structure, in which the location of the damage and defects can be revealed by the hot spots and/or cold spots. Due to the uncomplicated demonstration of measurement results, these images are also relatively easy to analyze. These advantages make IRT attractive in structural health monitoring and damage detection, especially when applied to large engineering structures where the instrumentation and measurement time required for other non-contact methods is usually considerably longer while the contact sensing—such as accelerometers and strain gauges—would miss large areas of the structure.

However, these easy-to-process high-definition images and videos come at a price. The requirements on storage space are one of the unavoidable problems of IRT, which is more significant when the safety requirement is critical so that online condition monitoring must be applied. In this case, a considerable amount of data needs to be stored to allow data traceability. Additionally, the demands on storage space have been increasing rapidly due to continuously advancing IR cameras, which are able to record data with higher resolution and frame rates.

In order to address this issue, data processing techniques such as feature extraction and data compression can be utilized.

There have been research outputs showing the possibility of using feature extraction on the thermal images acquired from IRT tests so that the images can be decomposed to underlying components containing different features [2] [3]. Through feature extraction, the components containing meaningful features such as the hot spots in the damaged regions can be extracted and preserved while the other components can be discarded in order to reduce the requirements on data storage [4].

However, the feature extraction process eliminates the temporal sequence of the data in exchange for the individual features, while in fact, the extracted results are not the identical representations of the original data at any specific time. For this reason, if the data at any specific time, or the time itself, is important, such as the time at which a critical event happens, the inclusion of feature extraction is not preferable.

A practical solution is to compress the raw data captured by IR cameras directly, which lowers the requirements for data storage from the outset while still maintaining the correct sequence of the data.

For the purpose of damage detection, Lugin et al. [5] developed a compression algorithm for data acquired from pulsed thermography tests, which was able to achieve a compression factor of up to 24–55 at a reproduction accuracy. Using medical infrared images as the vehicle, Schaefer et al. [6] compared several compression techniques, such as the Lossless JPEG, JPEG-LS, JPEG2000, PNG, and CALIC and demonstrated the possibility of compressing infrared images without losing critical information in medical applications.

However, research outputs demonstrating the viability of applying image compression to results from vibrothermographic tests are still limited. In vibrothermography, the results differ from the images captured in alternative IRT approaches in terms of several factors, which can add extra uncertainties to the compression process.

Firstly, during vibrothermographic tests, as external excitation is applied to the structure to induce vibration, the inspected object is not static during the tests. This phenomenon is particularly noticeable for vibration with high amplitude so that the motion of the inspected object can be reflected in the images acquired from vibrothermographic tests.

Secondly, in vibrothermographic tests, multiple hot spots may be generated, each from a different physical source. As a matter of fact, friction is not the only source of heat generation during a vibrothermographic test. There is also heat generated through other mechanisms such as plasticity-induced heat generation, viscoelastic heating and thermoelastic effect, although the frictional heat generation is often the most dominant source of heat generation in vibrothermographic tests [1] [7], which includes the experimental tests carried out in this research. However, even in situations where only frictional heat generation is considered, multiple hot spots always appear at different locations and different times. Among these hot spots, some are caused by the damage and defects while others may be caused by irrelevant features such as the less important interfaces. Typically, each hot spot often has a unique shape and amplitude. Specifically, damage and defects in regions with greater strain energy are more likely to develop higher frictional heat generation as the amount of heat generated through friction is determined by frictional force and slip rate (speed of relative motion) between the contact surfaces [1], which are positively correlated with local strain energy [3] [8]. The existence of multiple hot spots with varying amplitude across different frames of images is expected to present more challenges to the compression process.

Regarding the issue of additional irrelevant hot spots, a possible solution is to perform more customized excitation in order to induce the resonance of the damage and defects if there is sufficient knowledge of the damage and defects, which might provide more localized heat generation that can eliminate or reduce the magnitude of other hot spots experimentally. However, the prior knowledge of the characteristics of the potential damage and defects is not always available, in which case this approach is compromised.

Additionally, hot spots are likely to exhibit temperature fluctuations during measurement because of the inconstant heat generation rates during vibration caused by varying excitation force and other factors. Random fluctuations can also appear due to reasons such as camera noise, uneven emissivity and viewing angle of the IR camera.

Moreover, for in-field applications, reflections—including static and moving ones—may become unavoidable, which add extra noise to the measured results. Although image or pixel subtraction methods are often utilized to reduce reflections, removing reflections completely is still difficult. For this reason, in one of the experimental tests carried out in this research, it was not attempted to control the reflections in order to add extra challenges artificially, which should benefit the reliability and applicability of the conclusions obtained from this research.

Considering that data storage is an unavoidable problem of IRT in both laboratory tests and industrial applications, verifying the viability of applying image compression to results from vibrothermography, or other IRT approaches, appears to be necessary for improving its practicality in relevant applications.

In this research, data presented in the form of thermal images were acquired from two vibrothermographic tests on an aerospace-grade composite plate containing delamination. During the vibrothermographic tests, multiple hot spots generated through different mechanisms were developed. Specifically, in the second test, reflections were not eliminated which added extra noise to the measured results. After the experimental tests, the thermal images were firstly compressed to four different levels using an adaptive compression algorithm based on lifting discrete wavelet transform and set-partitioning embedded blocks [9] [10]. Five different methods, namely the compression ratio, mean squared error, peak signal-to-noise ratio, structural similarity index and coordinate modal assurance criterion, were applied to evaluate the performance of the compression process while identifying and locating the regions affected more significantly after image compression. Next, each set of data was processed through feature extraction, which separated the different hot spots so that the effects of the compression process on each individual hot spot could be assessed independently. The original images and the compressed data, as well as their respective extracted features, were compared to enable the consideration of the influence of the image compression process on the important features in the thermal images, such as the hot spots indicating the presence of damage, so that the viability of image compression in IRT, particularly vibrothermography, could be demonstrated.

Algorithm Overview and Theoretical Background

As explained in the previous section, the two main constituent components for data processing in this research are image compression and feature extraction.

In this research, the image compression process was achieved by an adaptive compression scheme [9] [10]. The compression algorithm is composed of a lifting discrete wavelet transform (LDWT) with set-partitioning embedded blocks (SPECK) that efficiently orders the wavelet coefficients by significance. The algorithm exploits the clustering of energy found in the transformed domain and concentrated on those areas of the set which have high energy. This allows those signals with higher information to be encoded first based on their energy content. The output of the SPECK module is the input to an optimized context-based arithmetic coder that generates the compressed bitstream.

Specifically, the advantages of the LDWT scheme are:

  1. 1.

    Lifting allows for adaptive wavelet transform;

  2. 2.

    The LDWT allows lossless reversible integer-to-integer transform; and

  3. 3.

    Lifting allows fast forward and inverse wavelet transforms and simple implementation.

The lifting scheme enables an adaptive wavelet transform, which means that the transform can start the analysis of a function from the most important (fundamental) layers and then build the detail layers by refining only the areas of interest. In the decoding part, the decoder can adaptively refine the detail information based on the requirements. An efficient wavelet coefficient sorting algorithm such as SPECK is helpful to sort subsets of wavelet coefficients by significance.

The SPECK uses three dynamic lists to store significant information of wavelet coefficients for compression purposes, which are the list of significant pixels (LSP), list of insignificant pixels (LIP) and list of insignificant sets (LIS). To check the significance, the algorithm compares the wavelet coefficients in a set with respect to a sequence of decreasing threshold. In case that the wavelet coefficient has its magnitude above the threshold, the set is split into four subsets whose significance with respect to the threshold is tested, otherwise, the set is not split. Finally, the algorithm rescans those pixels found to be significant with respect to previous thresholds and refines their magnitude. The function used for testing the set significant state is defined by the following formula:

$$ {S}_i\left(\Gamma \right)=\left\{\begin{array}{c}1,\kern0.5em {2}^i\le \underset{\left(m,n\right)\in \Gamma}{\max}\left\{\left|{c}_{m,n}\right|\right\}\le {2}^{i+1}\\ {}0,\kern0.5em \mathrm{otherwise}\end{array}\right. $$
(1)

and

$$ i=\left\lfloor {\log}_2\left({\max}_{\left(m,n\right)}\left\{\left|{c}_{m,n}\right|\right\}\right)\right\rfloor, $$
(2)

where cm, n is the coefficient value for position (m, n) in the wavelet sub-bands. The Γ indicates the set of coefficients and Si(Γ) is used for the significant state of the set Γ, where 1 indicates the set is significant to the threshold T = 2i, and 0 indicates it is insignificant and a bit will be emitted for the entire insignificant set. The insignificant pixels and insignificant sets are put in LIP and LIS, respectively; the significant pixels are put in the LSP.

The context modelling in SPECK has been achieved by arithmetic coding (AC), which can obtain optimal performance thanks to its ability to represent information with fractional bits. AC has been widely utilized by various image and video codecs, such as the QM coder in JPEG [11], the MQ coder in JPEG2000 [12], and the context-based adaptive binary arithmetic coder (CABAC) in H.264/AVC and HEVC [13]. The context modelling provides estimates of the conditional probabilities of the coding coefficients and exploits the redundancy present in the neighborhood of the current coefficient to encode.

The arithmetic coder employed in this work is the MQ-coder as specified in the JBIG and JPEG2000 standard [12] [14]. The MQ-coder is a binary alphabet adaptive multiplication-free arithmetic coder. A probability model has been implemented as a 47-state Finite State Machine (FSM) for the coding process. In the FSM, each state contains coding information, a set of probability mapping rules (look-up table) which are used to interpret and manipulate the context state associated with the current coding context [14]. This approach can estimate the probability of the context more efficiently and take into consideration the non-stationary characteristic of the symbol string. The MQ-coder can adapt to the input bitstream by estimating the probability through the use of two tables: Dynamic table and Static table. The Dynamic table contains the information of each context, and the Static table describes the probability state transition for MQ-coder [14].

In this research, the experimental results indicated that the data compressor was able to efficiently compress the signals from the vibrothermographic tests with high-quality reconstruction as it will be shown later in this article. The compression ratio (CR) was utilized to measure the relative reduction in the size of data after the image compression process, which can be easily calculated as follows:

$$ \mathrm{CR}=\frac{L_{\mathrm{orig}.}}{L_{\mathrm{comp}.}}, $$
(3)

where Lorig. and Lcomp. are the sizes of the original data and the compressed output respectively.

In addition to the overall compression ratio, for the purpose of damage detection and structural health monitoring, it must be verified that the regions that contained valuable information and meaningful features such as the hot spots were less affected during the compression process as the compression mostly influenced the areas without meaningful features. In order to achieve this objective, after the image compression process, three different methods were utilized to identify and locate the differences between the original images and the compressed data. Specifically, these three techniques are the mean squared error (MSE), structural similarity (SSIM) index and coordinate modal assurance criterion (COMAC).

Among these three methods, the MSE provides the simplest approach. Firstly, the MSE between two images I and K which have the same pixel resolution M × N can be calculated from:

$$ \mathrm{MSE}=\frac{1}{MN}{\sum}_{m=1}^M{\sum}_{n=1}^N{\left[I\left(m,n\right)-K\left(m,n\right)\right]}^2 $$
(4)

where I(m, n) is the intensity of the pixel (m, n) in the first image while K(m, n) is the intensity of the same pixel in the second image.

Now if I represents a set of original images that have not been compressed while K represents the corresponding set of images after image compression, the MSE for each pixel between every image pair can be calculated using:

$$ \mathrm{MSE}\left({m}_i,{n}_i\right)={\left(I\left({m}_i,{n}_i\right)-K\left({m}_i,{n}_i\right)\right)}^2, $$
(5)

where, for the specific application in this research, I(mi, ni) is the intensity of the pixel (m, n) in the ith image of the original data set and K(mi, ni) is the intensity of the same pixel in the corresponding image of the compressed data set. After calculating the MSE for every image pair, the average MSE for the pixel (m, n) between the two sets of images can be calculated as follows:

$$ {\mathrm{MSE}}_{\mathrm{avg}.}\left(m,n\right)=\frac{1}{J}{\sum}_{i=1}^J\mathrm{MSE}\left({m}_i,{n}_i\right). $$
(6)

where it has been assumed that each set contains J images. The calculated average MSE map can then be utilized to visualize the pixel-wise differences between the two sets of images. Eventually, a single MSE value can be calculated to evaluate the overall difference between the two sets of images using:

$$ \mathrm{MSE}=\frac{1}{MN}{\sum}_{m=1}^M{\sum}_{n=1}^N{\mathrm{MSE}}_{\mathrm{avg}.}\left(m,n\right) $$
(7)

Based on the overall MSE, the peak signal-to-noise ratio (PSNR) can be calculated between the original and the compressed data from:

$$ \mathrm{PSNR}=10{\log}_{10}\left(\frac{{\operatorname{MAX}}_I^2}{\mathrm{MSE}}\right), $$
(8)

where MSE is the overall mean squared error between the original images I and the compressed images K. MAXI is the maximum possible pixel value of the image, which is 255 in this case.

Compared to the simple and straightforward methods such as the MSE and PSNR, the SSIM was designed as an improvement by considering three components, which are the luminance l(x, y), the contrast c(x, y) and the structure s(x, y) [15], so that:

$$ l\left(x,y\right)=\frac{2{\mu}_x{\mu}_y+{C}_1}{\mu_x^2+{\mu}_y^2+{C}_1}, $$
(9)
$$ c\left(x,y\right)=\frac{2{\sigma}_x{\sigma}_y+{C}_2}{\sigma_x^2+{\sigma}_y^2+{C}_2} $$
(10)

and

$$ s\left(x,y\right)=\frac{\sigma_{xy}+{C}_3}{\sigma_x{\sigma}_y+{C}_3}, $$
(11)

where μx is the mean intensity of a window x, μy is the mean intensity of another window y, \( {\sigma}_x^2 \) is the variance of x, \( {\sigma}_y^2 \) is the variance of y and σxy is the covariance of x and y. The constants C1, C2 and C3 are included to avoid instability when the denominators are close to zero [15]. The constants can be calculated using:

$$ {C}_1={\left({K}_1L\right)}^2, $$
(12)
$$ {C}_2={\left({K}_2L\right)}^2 $$
(13)

and

$$ {C}_3={\left({K}_3L\right)}^2, $$
(14)

where L is the dynamic range of the pixel values while K1, K2 and K3 are arbitrary small constants that are much less than 1 [15].

By combining the three components, the SSIM index can be calculated as follows:

$$ \mathrm{SSIM}\left(x,y\right)={\left[l\left(x,y\right)\right]}^{\alpha}\cdotp {\left[c\left(x,y\right)\right]}^{\beta}\cdotp {\left[s\left(x,y\right)\right]}^{\gamma }. $$
(15)

By setting α = β = γ = 1 and C3 = C2/2, it can be simplified to:

$$ \mathrm{SSIM}\left(x,y\right)=\frac{\left(2{\mu}_x{\mu}_y+{C}_1\right)\left(2{\sigma}_{xy}+{C}_2\right)}{\left({\mu}_x^2+{\mu}_y^2+{C}_1\right)\left({\sigma}_x^2+{\sigma}_y^2+{C}_2\right)}. $$
(16)

In this research, x was taken from an image of the original data set while y was the counterpart from the corresponding image of the compressed data set. After calculating the local SSIM values between every image pair, the average local SSIM for a window m between the two sets of images can be obtained from:

$$ {\mathrm{SSIM}}_{\mathrm{avg}.}\left({x}_m,{y}_m\right)={\sum}_{i=1}^J\mathrm{SSIM}\left({x}_{mi},{y}_{mi}\right). $$
(17)

where it has been assumed that each set contains J images.

Similar to the MSE, a single SSIM value can be calculated to represent the overall similarity between two images I and K. This was defined as the mean SSIM (MSSIM) [15]. By assuming that there are M windows in each image, it can be calculated from:

$$ \mathrm{MSSIM}\left(I,K\right)=\frac{1}{M}{\sum}_{m=1}^M\mathrm{SSIM}\left({x}_m,{y}_m\right). $$
(18)

This can then be extended to two sets of images by substituting the average local SSIM between the two sets of images, which is the SSIMavg. from Eq. 17, for the SSIM in Eq. 18, so that a single SSIM value between the two sets of images can be calculated. Additionally, the results calculated from Eq. 17 can be utilized to construct a pixel-by-pixel map visualizing the influence from the image compression process, where the details can be found in the article by Wang et al. [15]. In the SSIM map, regions with smaller local SSIM values correspond to areas that have been affected more significantly through the image compression process.

Eventually, the coordinate modal assurance criterion (COMAC), originally developed to identify the degrees-of-freedom that cause differences between two sets of mode shapes [16], was utilized as the third method to locate the differences between the original and the compressed images. The COMAC can be calculated from:

$$ \mathrm{COMAC}(m)=\frac{{\left({\sum}_{i=1}^J\left|{\left({\varphi}_A\right)}_{mi}{\left({\varphi}_X\right)}_{mi}^{\ast}\right|\right)}^2}{\sum_{i=1}^J{\left({\varphi}_A\right)}_{mi}^2{\sum}_{i=1}^J{\left({\varphi}_X\right)}_{mi}^{\ast 2}}, $$
(19)

where, for the specific application in this research, (φA)mi represents the intensity of the mth pixel in the ith image of the first data set. (φX)mi is the intensity of the same pixel in the corresponding image in the second data set and J is the total number of correlated image pairs. The COMAC, which is a vector, can be converted back to a matrix which has the same dimension as the pixel resolution of the original images and every element in COMAC can be placed back at their original pixel position. The value of each element in the COMAC vector can be used to define the height and/or the color of each pixel in the re-created image so that the pixels causing the differences between two sets of images can be located easily. The COMAC is normalized so always has a value between 0 and 1. A larger COMAC value indicates higher similarities at a specific element, which is a pixel in this case, between the two sets of data.

In this research, the MSE, SSIM and COMAC maps were calculated between the original and the compressed sets of data. The regions affected more significantly in the image compression process would have higher MSE values but lower SSIM and COMAC values than those affected less. Based on this information, the areas influenced more significantly through the image compression process could be identified and located clearly, with which the effectiveness of the image compression could be appraised.

To complete the process, the original and the compressed data sets were further processed through feature extraction. Feature extraction was performed after the completion of image compression, which attempted to separate the underlying features in the thermal images such as the hot spots. With the different features separated, the effects of the compression process on each individual hot spot could be assessed independently.

In feature extraction, it is assumed that each underlying feature is generated by a unique physical source. For the case in this research, these unique features were mixed to form the thermal images that were captured by the IR camera. The mixing process can be described as:

$$ X= AS, $$
(20)

where X is the observed mixtures, which were the thermal images in this research, S is the source signals, which were the underlying features such as the hot spots generated by different physical sources, and A is the matrix that performs the mixing process.

The feature extraction can be considered as a reverse of the mixing process, so that:

$$ Y= WX, $$
(21)

where Y contains the separated features. The objective of a feature extraction process is to calculate a proper W so that the extracted signals Y are as close to the original signals S as possible. These components in Y are assumed to be mutually uncorrelated (as in the principal component analysis (PCA) and factor analysis (FA)) [4] or statistically independent (as in the independent component analysis (ICA)) [17]. W is the matrix that performs the unmixing process, which needs to be determined in order to calculate the correct components from the feature extraction.

Multiple feature extraction algorithms can achieve this objective of calculating the unmixing matrix W or its constituting weight vectors wi, among which PCA, FA and ICA are three of the most common techniques [17]. Among these three techniques, ICA, specifically, FastICA [18], was used in this research, as it has been verified on the thermal images used in this research that it was able to separate the underlying features successfully while still maintaining their correct characteristics, which was unable to be achieved by PCA [3].

However, despite being able to provide more robust results than alternative techniques in most situations, FastICA has certain limitations that must be considered.

Firstly, due to the calculation algorithm of FastICA, the number of extracted components is equal to the number of input signals provided, given that no dimensionality reduction is performed. In most cases, the number of extracted components cannot reflect the actual number of underlying features.

Another limitation of FastICA that is relevant to IRT is that FastICA cannot determine the sign or scale of each component, which means the heat generated by some heat sources may appear as cold spots in the components extracted by FastICA due to the change in sign [19]. For this reason, the extracted components in this research were further processed so that the hot spots always appear as red in the generated images of the extracted components. Another consequence of this limitation is that the values of the elements in the components extracted by FastICA are not equivalent to the original temperature data due to the change in scale, although the scale does not affect the relative magnitude of data across pixels in each extracted component.

Another characteristic of FastICA to be considered is that the order of the extracted components is not determined by any meaningful criterion, although this can be solved by introducing additional calculations, such as using kurtosis, to re-order the extracted components, which has been demonstrated by the authors [3].

Experimental Development and Data Acquisition

As explained in the section “Introduction”, the experimental data were acquired from two vibrothermographic tests performed on an aerospace-grade composite plate containing delamination. The plate was made of HexPly® 8552 AS4 epoxy matrix woven carbon prepregs composite material which is commonly utilized in primary aerospace structures [20]. The damage considered was vertical delamination created in the central region of the plate through dynamic fatigue.

Composite materials are used more and more routinely in industries due to their significant advantages in physical properties in terms of strength-to-weight ratio when compared to traditional materials such as metals. However, the unique structure of ply-based composite materials has brought in new damage mechanisms such as delamination. Delamination is arguably the most common and most challenging type of damage in composite materials used in aerospace-related applications. However, this type of damage can be difficult to detect as they are rarely observable on the surface of the material. For IRT specifically, the hot spots generated due to delamination are also more difficult for IR cameras to capture than those generated by visible defects. These reasons justify the attention that composite materials have received in research activities involving IRT [21] [22] [23] [24] [25] [26]. Because of the properties of the structure being inspected in this research and the type of damage, vibrothermography was more suitable than the conventional heat-based IRT approaches thanks to the friction-based heat generation mechanism.

As explained in the section “Introduction”, frictional heat generation rate during vibrothermographic tests is positively correlated with local strain energy. For a vibrating structure, high strain energy is achieved around resonance, at which response of the structure is maximized. For this reason, the calculation of strain energy distribution in the structure can be achieved by a modal analysis.

Specifically, modal analysis is able to numerically calculate and experimentally measure the natural frequencies (which are approximately equal to the resonant frequencies for structures with relatively low damping) as well as the mode shape and the associated strain energy distribution of each mode of vibration. When a structure is in an operating condition, its deflection shape is a combination of all excited mode shapes. For this reason, calculating the strain energy map of each mode of vibration is able to predict the actual strain energy distribution in the structure [27].

For active vibrothermographic testing where external excitation is applied to the structure to induce vibration, with the necessary knowledge of the distribution of strain energy at each resonance, parameters of the excitation force can be selected to target a specific mode of vibration in order to ensure high strain energy concentration in the interested regions during the structure’s vibration.

For this reason, in this research, in order to perform the vibrothermographic test effectively and efficiently, a modal test was performed first to calculate the modal parameters of the structure so that the proper excitation frequency and location could be selected. The rig set-up and the flowchart demonstrating the procedures of the experimental tests are shown in Figs. 1 and 2. During all tests, the composite plate was suspended using two springs with elastic bands and connected to an LDS V201 electrodynamic shaker through a pushing rod which was screwed into a connector glued onto the back of the plate, as shown in Fig. 3. As shown in Fig. 3 (a), multiple reflective tapes were glued onto the plate to allow for the measurement using a scanning laser Doppler vibrometer during the modal test. All reflective tapes but one were then removed before the vibrothermographic tests.

Fig. 1
figure 1

The rig for the experimental tests

Fig. 2
figure 2

The flowchart of the experimental tests

Fig. 3
figure 3

a The suspended plate and b the attachment point of the electrodynamic shaker

After the completion of the modal analysis, the first mode of vibration was selected for the vibrothermographic test. The high-definition strain energy distribution in the structure during this mode of vibration was calculated from an updated finite element model. The strain energy distribution was superimposed on its corresponding mode shape and shown in Fig. 4.

Fig. 4
figure 4

Strain energy distribution of the first mode of vibration

As shown in Fig. 4, the central region of the plate where the damage was located had high strain energy concentration in the first mode of vibration, which demonstrated its suitability for vibrothermographic testing.

During the vibrothermographic test, the electrodynamic shaker was set to provide a sinusoidal excitation force with low amplitude. The excitation frequency was set to match the natural frequency of the first mode of vibration of the composite plate, which was experimentally measured as 119 Hz. The location of excitation was also adjusted in the vibrothermographic tests due to the limitations from the aspect ratio of the IR camera. As mentioned previously, all reflective tapes except for the one at the excitation point were removed for the vibrothermographic tests. The remaining reflective tape was left to enable the measurement of the driving point frequency response function which was used to ensure that the natural frequencies were not noticeably altered after changing the location of the excitation point.

A FLIR T650sc IR camera was used in the vibrothermographic tests for data acquisition. The camera has resolution of 640 × 480 and thermal sensitivity of 0.02 °C, which means it was able to capture the subtle temperature changes in the inspected structure during the vibrothermographic tests.

As mentioned previously, two experimental vibrothermographic tests were performed. In the first test, barriers were used to block the light to control the reflections on the composite plate, while in the second test the IR camera was placed directly in front of the plate, in which case the reflections were significantly stronger.

In both tests, data recording had been started before the shaker was turned on in order to capture the initial temperature of the plate. After five minutes, the electrodynamic shaker was turned on to provide the excitation force. Due to the frequency and location selected for the excitation force, the deflection shape of the plate closely resembled the mode shape of its first mode of vibration. During the test, heat was generated in the damaged region because of the friction caused by the vibration. The frictional heat generation was able to be captured by the IR camera so that a hot spot could be observed in the damaged region. A separate hot spot was also developed at the excitation point primarily due to the heat transferred from the shaker into the plate through the pushing rod. At the end of the active testing, the shaker was turned off, so the temperature of the plate started to decrease, during which the heat concentration at the excitation point and in the damaged region gradually dissipated. In both tests, data were recorded at 15-s intervals. The first test lasted 45 min, during which 181 frames of data were recorded. The second test lasted 50 min, during which 200 frames of data were recorded. The timelines of the two vibrothermographic tests are described in Tables 1 and 2 respectively.

Table 1 Timeline of the first vibrothermographic test
Table 2 Timeline of the second vibrothermographic test

As shown in Fig. 5 (a), the maximum temperature variance across the plate was less than 0.2 °C before the electrodynamic shaker was turned on. After a 30-s excitation, two separate hot spots could be observed clearly in the damaged region and at the excitation point respectively, as shown in Fig. 5 (b). Fig. 5 (c) shows that the heat concentration in the damaged region dissipated swiftly within 15 s after the shaker was turned off, while it took 20 min (Fig. 5 (d)) for the hot spot at the excitation point to be fully dissipated. The reason could be mostly attributed to the heat generated in the shaker, which was significantly higher than the heat generated in the plate, continued to be transferred into the plate after the shaker was turned off.

Fig. 5
figure 5

Photographs taken at (a) 0:30, (b) 5:30, (c) 25:15 and (d) 45:00 of the first vibrothermographic test

Within the captured images, it can be observed that multiple concerns regarding the results from vibrothermographic tests that would create additional uncertainties to the image compression process, which are summarized and explained in section “Introduction”, have been confirmed.

Firstly, there were two main hot spots on the plate, one generated by friction and the other caused primarily by heat transfer. There were additional hot spots in the background, as shown in Fig. 5 (c) and Fig. 5 (d). The two main hot spots started to appear at different time, and the peak amplitude was also different. Specifically, the frictional heat generation started more rapidly, but the continuous heat transfer from the shaker was able to eventually accumulate a more significant amount of energy. Color anomalies and fluctuations caused by unwanted factors, such as camera noise, uneven emissivity and viewing angle, could also be observed clearly as shown in Fig. 5 (a), which was taken before the shaker was started.

In addition to the concerns mentioned above, reflections were included in the second test, where barriers were removed, and the camera was placed directly in front of the plate.

The results from the second test were considerably more complicated due to the additional reflections caused by the removal of the barrier. Specifically, the reflections were predominantly caused by the camera which was placed directly in front of the plate in this test. The reflections were captured clearly as shown in Fig. 6 (a). In the first test, it only took four seconds for a hot spot to emerge in the damaged region while in this test the heat generated in the damaged region was significantly less noticeable because it was overshadowed by the strong reflections (Fig. 6 (b)). Fig. 6 (c) and Fig. 6 (d) show that an obvious hot spot was developed at the excitation point because of its distance from the reflections. Despite not being able to observe the hot spot in the damaged region directly, the more localized thermal pattern in the central region of the plate, as shown in Fig. 6 (c) and Fig. 6 (d), may be a hint of the heat generated by the friction in the damaged region. Similar to the first test, it again took a considerable amount of time for the hot spot at the excitation point to disappear fully because of the significant amount of energy accumulation due to heat transfer.

Fig. 6
figure 6

Photographs taken at (a) 0:30, (b) 6:00, (c) 20:00, (d) 35:00, (e) 40:00 and (f) 50:00 of the second vibrothermographic test

The data recorded during the vibrothermographic tests have confirmed that the friction in the damaged region was able to cause detectable local heat generation. Apart from the main feature, other relatively minor features were also developed and captured in the images, such as the irrelevant hot spots, color fluctuations and reflections, which would add more challenges to the compression process.

For the next step, these images were compressed into four different levels. The CR, MSE, PSNR, SSIM and COMAC were then utilized to compare the original and the compressed images. Next, feature extraction through FastICA was applied to investigate the differences between the extracted components and thus to study the effects of the compression process on the characteristics of each important feature.

Data Processing

Image Compression

The data were compressed into four levels (named as C0, C1, C2 and C3) using the adaptive compression algorithm described in section “Algorithm overview and theoretical background”, where the data set C0 had the lowest compression ratio while C3 had the highest compression ratio among these four sets of compressed images.

After the compression process, the methods introduced in the section “Algorithm overview and theoretical background” were applied to evaluate the performance of the image compression process. The results between the original data set and each compressed data set in addition to the basic information of the data and the exact space required for storage of each data set are summarized in Tables 3 and 4. The compressed versions of Figs. 5 (b) and 6 (c) are shown in Fig. 7 and Fig. 8 respectively as examples, in which new colormaps were created and applied based on the jet color scheme in MATLAB.

Table 3 Compression results, the first vibrothermographic test
Table 4 Compression results, the second vibrothermographic test
Fig. 7
figure 7

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of Fig. 5 (b)

Fig. 8
figure 8

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of Fig. 6 (c)

The values in Tables 3 and 4 demonstrate that the results from all five methods, namely CR, MSE, PSNR, SSIM and COMAC, changed monotonically as the level of compression increased. To be specific, the CR and MSE increased for data with a higher level of compression while the PSNR, SSIM and COMAC all decreased. However, the exact values and the relative changes calculated from each method are not comparable because of the different calculation algorithms of the methods.

By observing the images in Figs. 7 and 8 directly, the characteristics of the major features were relatively unchanged after the image compression. Although it was noticeable that the images had been compressed, the significance of the image compression in different regions of the images was unclear. In order to determine the effects of the image compression process in different regions of the images, the maps of MSE, SSIM and COMAC, which have been explained in the section “Algorithm overview and theoretical background”, were calculated between the original data set the data set with the highest compression ratio (C3) and displayed in Figs. 9 and 10. Specifically, the (1-SSIM) and (1-COMAC) maps are displayed so that the regions affected more significantly during the image compression process will appear in red in the generated images.

Fig. 9
figure 9

The (a) MSE, (b) (1-SSIM) and (c) (1-COMAC) maps between the original data set and the compressed data set C3, the first vibrothermographic test

Fig. 10
figure 10

The (a) MSE, (b) (1-SSIM) and (c) (1-COMAC) maps between the original data set and the compressed data set C3, the second vibrothermographic test

By observing the images in Figs. 9 and 10, although the maps from the three different methods have dissimilar values and scales because of their different calculation algorithms, it is ascertainable that the regions with greater differences after the image compression process were mostly the areas that did not contain meaningful features, such as the background regions in Fig. 9 and the boundary regions in Fig. 10.

This observation has corroborated the possibility of applying image compression to results from vibrothermography. In vibrothermography or most IRT approaches in general, the hot spots in the damaged regions are likely to change more drastically over time while the irrelevant features in the background will stay relatively unchanged. The high level of activity of the hot spots generated by the meaningful sources means they will remain relatively unaffected during the image compression process. Conversely, as the features developed by the irrelevant background sources and the reflections are usually static during the measurement, these regions will be affected more significantly during image compression.

Based on the observations described above, it is reasonable to assert that the meaningful features would still be successfully extracted after feature extraction even though the image compression had been performed, and the characteristics of the features would remain mostly unchanged.

Feature Extraction

After acquiring the four sets of compressed versions of the results from each vibrothermographic test, feature extraction was performed on all ten sets of data using the FastICA described in section “Algorithm overview and theoretical background”. Due to the nature of FastICA that has been explained, the number of extracted components was equal to the number of signals provided, which means that there was a large number of extracted components that did not contain any actual meaningful feature. For this reason, the extracted components were reviewed manually, after which two meaningful components in each data set were selected, so that the effects of the image compression process on each meaningful feature could be assessed independently.

In the first data set, the two meaningful features were the hot spots generated in the damaged region and at the excitation point, which are shown in Figs. 11 and 12 respectively.

Fig. 11
figure 11

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the extracted hot spot in the damaged region, the first vibrothermographic test

Fig. 12
figure 12

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the extracted hot spot at the excitation point, the first vibrothermographic test

It is demonstrated in Figs. 11 and 12 that the two important features were extracted successfully despite image compression being performed. The differences between the results from the uncompressed and compressed images are only distinguishable under close inspection. Specifically, the compressed data sets were still able to provide features with correct shape and location. However, the image compression process had marginally influenced the size of the features. By observing closely, the features extracted from images sets with higher compression ratio tended to have a smaller size, which was true for both features.

To explain this, the edge regions of the hot spot always experience less drastic changes over time during the vibrothermographic tests. This means the edge regions were likely to be compressed more significantly, and the size of the area being affected was determined by the level of compression. In this scenario, when the compression ratio increased, a larger percentage of the hot spot would be affected, which correspondingly reduced the size of the remaining feature.

In order to quantify the effects of image compression on the size of the hot spots, each extracted image was normalized to have zero mean and unit variance. If the intensity values of pixels were considered as a probability distribution, pixels constituting the hot spots must lie in the upper-end region of the distribution. For this reason, it was assumed that elements with values greater than 3 were parts of the hot spots, where the value—3—used here was determined after calibrations on this specific data set. The number of elements (pixels) matching this requirement in each extracted component is summarized in Table 5. Pixels not matching this requirement were filtered out. The reconstructed images after the filtering process are shown in Figs. 13 and 14.

Table 5 The number of elements (pixels) that each feature consists of, the first vibrothermographic test
Fig. 13
figure 13

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the reconstructed hot spot in the damaged region, the first vibrothermographic test

Fig. 14
figure 14

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the reconstructed hot spot at the excitation point, the first vibrothermographic test

Based on the results in Table 5, Figs. 13 and 14, it can be verified that the image compression has altered the size of the features while the shape and location of the feature remained mostly unchanged. The reduction in the size of the features was positively correlated with the compression ratio. The changes on the size of the meaningful features were relatively insignificant considering the high compression ratios, as it has been confirmed through the results from MSE, SSIM and COMAC that the compression process mostly affected the less important areas containing irrelevant features. Specifically, for a compression that could save 97% of storage space, the reduction of the size of the meaningful features was only less than 10%.

In the second data set, apart from the hot spots in the damaged region and at the excitation point, there were additional features which were the reflections on the plate due to removal of the barrier.

FastICA was still able to separate the hot spots in the damaged region (Fig. 15) and at the excitation point (Fig. 16) from the irrelevant thermal patterns successfully, even though the reflections covered a large portion of the images. However, the thermal pattern of the reflections on the reflective tape was so significant that an independent component containing only this feature was extracted, which caused this area to be removed from the feature of the excitation point, as shown in Fig. 16.

By observing the extracted features, it can be inferred that the observations from the first data set still apply and so the same conclusions can still be made. The compressed data were still able to provide correctly identified features. The size of the features decreased more significantly as the compression ratio increased, which has been verified in Table 6, Figs. 17 and 18 using the same technique applied to the results from the first test.

Table 6 The number of elements (pixels) that each feature consists of, the second vibrothermographic test

Additionally, in this data set, the hot spot caused by the frictional heat generation in the damaged region was partially mixed with the reflections, as shown in Fig. 15. The image compression process had moderately affected the regions containing the reflections, as shown in Fig. 10, which caused significantly more obvious differences between Fig. 15 (a) and Fig. 15 (b), Fig. 15 (c), Fig. 15 (d) and Fig. 15 (e) compared to the cases in other extracted components. This behavior is clearly reflected in Table 6 and Fig. 17 as the percentage decreases in the number of elements (pixels) from the original set to C0, C1, C2 and C3 are significantly higher compared to the values in the other three features. For this feature, the number of elements (pixels) was reduced by 66.81% from the original set to C3 while this number was only −21.80%, −24.91% and − 23.81% for the other three features respectively.

Fig. 15
figure 15

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the extracted hot spot in the damaged region, the second vibrothermographic test

Fig. 16
figure 16

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the extracted hot spot at the excitation point, the second vibrothermographic test

Fig. 17
figure 17

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the reconstructed hot spot in the damaged region, the second vibrothermographic test

Fig. 18
figure 18

The (a) original, (b) C0, (c) C1, (d) C2 and (e) C3 versions of the reconstructed hot spot at the excitation point, the second vibrothermographic test

Discussion and Conclusion

In this article, research outputs focusing on the verification of the viability of applying image compression to results from vibrothermographic tests is presented.

It has been shown through results from two tests that the compressed images were still able to preserve the meaningful features, such as the hot spot in the damaged region, with overall accurate characteristics despite the challenges caused by the irrelevant hot spots, fluctuations and reflections in the images.

By using three different methods, namely MSE, SSIM and COMAC, it has been verified that the image compression process was able to have more significant influences in regions mostly consisting of less-important features that did not change drastically across frames in measurement results while having a considerably smaller impact in regions with important time-variant features. By observing the extracted components from the original and the compressed images, it has been discovered that the compression process reduced the size of the important features slightly by compressing their outer regions. The size of the area being removed was positively correlated with the compression ratio, where a larger compression ratio always led to hot spots with smaller size. However, the influences on the meaningful features were still significantly smaller than the impact in the less-important regions. Specifically, a compression process which was able to save 97% of storage space only altered the size of the meaningful features by less than 10% in this research.

For this reason, a sacrifice of the accurate determination of the size of a feature can be highly beneficial in exchange for the storage space saved from image compression when the requirement on the accurate determination of the size of the feature is not critical. The hot spots extracted from the compressed data are still able to maintain generally accurate characteristics such as the location and shape.

However, it must be noted that in the thermal images used in this research, the regions with valuable information only occupied relatively small areas. For images with larger areas covered by valuable information, it can be expected that a comparably high compression ratio will not be equally viable because it will potentially lead to a greater reduction in terms of the size of the meaningful features after the compression process.

Aside from the observations and conclusions obtained from this research, there are still potential improvements that can be made.

Firstly, it is possible to use other image compression methods to verify if the same phenomena still occur, which could increase the generality of the conclusions obtained from this research.

Secondly, the structure used in this research is still relatively simple, in which only several features were involved. It can be beneficial to repeat the procedures in this research on a more complicated structure with more complex geometry, which could increase the practicality of this approach.

Lastly, although the MSE, SSIM and COMAC were able to provide similar results in terms of the influences from the image compression process in different regions of the images, minor disagreements could still be observed, such as the bottom-left region in both Figs. 9 and 10. The differences have been attributed to the dissimilar calculation algorithm of each method in this article, but it might be worthwhile to look for more specific reasons.

Research Data Accessibility

The data produced during this research project and used in this article have been uploaded to Google Drive, which can be accessed at https://drive.google.com/drive/folders/1jEwtYENME9P9qh4isZeS2dTQGZyBvl-n?usp=sharing.