Acessibilidade / Reportar erro

Contact Failure Identification in Multilayered Media via Artificial Neural Networks and Autoencoders

Abstract

The estimation of defects positioning occurring in the interface between different materials is performed by using an artificial neural network modeled as an inverse heat conduction problem. Identifying contact failures in the bonding process of different materials is crucial in many engineering applications, ranging from manufacturing, preventive inspection and even failure diagnosis. This can be modeled as an inverse heat conduction problem in multilayered media, where thermography temperature measurements from an exposed surface of the media are available. This work solves this inverse problem with an artificial neural network that receives these experimental data as input and outputs the thermalphysical properties of the adhesive layer, where defects can occur. An autoencoder is used to reduce the dimension of the transient 1D thermography data, where its latent space represents the experimental data in a lower dimension, then these reduced data are used as input to a fully connected multilayer perceptron network. Results indicate that this is a promising approach due to the good accuracy and low computational cost observed. In addition, by including different noise levels within a defined range in the training process, the network can generalize the experimental data input and estimate the positioning of defects with similar quality.

Key words
Inverse heat conduction problem; artificial neural networks; contact failure identification; applied artificial intelligence; denoising autoencoder

Introduction

Heat transfer in multilayered media is a phenomenon that can be fundamentally important in different applications of science and engineering, such as manufacturing of composite materials (Danes et al. 2003DANES F, GARNIER B & DUPUIS T. 2003. Predicting, measuring, and tailoring the transverse thermal conductivity of composites from polymer matrix and metal filler. Int J Thermophys 24(3): 771–784. doi:10.1023/A:1024096401779.
10.1023/A:1024096401779...
), oil and gas industry (Su et al. 2012SU J, CERQUEIRA DR & WANG S-Z. 2012. Thermal design of multi-layered composite pipelines for deep water oil and gas production. Int J Comp Appl Technol 43(3): 248–259. doi:10.1504/IJCAT.2012.046311.
10.1504/IJCAT.2012.046311...
), thermal based health treatments (Sarkar et al. 2015SARKAR D, HAJI-SHEIKH A & JAIN A. 2015. Temperature distribution in multi-layer skin tissue in presence of a tumor. Int J Heat Mass Transf 91: 602–610. doi:10.1016/j.ijheatmasstransfer.2015.07.089.
10.1016/j.ijheatmasstransfer.2015.07.089...
), aerospace engineering (Jayaseelan et al. 2015JAYASEELAN DD, XIN Y, VANDEPERRE L, BROWN P & LEE WE. 2015. Development of multi-layered thermal protection system (TPS) for aerospace applications. Comp Part B Eng 79: 392–405. doi:10.1016/j.compositesb.2015.04.017.
10.1016/j.compositesb.2015.04.017...
) and many others. More specifically, defect identification in the process of bonding among different materials can be formulated as a multilayered media and solved as an Inverse Heat Conduction Problem (IHCP), where estimating the positioning of air gaps, oil bubbles, and cracks, usually is performed by estimating thermophysical properties of the adhesive layer. Therefore, estimating heterogeneity in this layer may be crucial to evaluate the bonding quality (Abreu et al. 2014ABREU LAS, ORLANDE HRB, KAIPIO J, KOLEHMAINEN V, COTTA RM & QUARESMA JNN. 2014. Identification of contact failures in multilayered composites with the Markov chain Monte Carlo method. J Heat Transf 136(10). doi:10.1115/1.4027364.
10.1115/1.4027364...
, 2016ABREU LAS, COLAÇO MJ, ORLANDE HRB & ALVES CJS. 2016. Thermography detection of contact failures in double layered materials using the reciprocity functional approach. Appl Therm Eng 100: 1173–1178. doi:10.1016/j.applthermaleng.2016.02.078.
10.1016/j.applthermaleng.2016.02.078...
, Mascouto et al. 2020MASCOUTO FS, KNUPP DC, ABREU LAS, SANCHES EL & SPHAIER LA. 2020. Detection of contact failures employing combination of integral transforms with single-domain formulation, finite differences, and Bayesian inference. Num Heat Transf Part A Appl 77(6): 599–618. doi:10.1080/10407782.2020.1713637.
10.1080/10407782.2020.1713637...
).

In classical approaches, IHCPs can be treated as optimization problems, where a least squares functional must be minimized (Beck et al. 1985BECK JV, BLACKWELL B & CLAIR JR CRS. 1985. Inverse heat conduction: Ill-posed problems. John Wiley & Sons. , Moura Neto & Silva Neto 2013MOURA NETO FD & SILVA NETO AJ. 2013. An introduction to inverse problems with applications. Springer Science & Business Media. doi:10.1007/978-3-642-32557-1.
10.1007/978-3-642-32557-1...
), or as bayesian inference problems, such as the Maximum a Posteriori and the Markov Chain Monte Carlo methods (Kaipio & Fox 2011KAIPIO JP & FOX C. 2011. The Bayesian framework for inverse problems in heat transfer. Heat Transf Eng 32(9): 718–753. doi:10.1080/01457632.2011.525137.
10.1080/01457632.2011.525137...
, Orlande 2012ORLANDE HRB. 2012. Inverse problems in heat transfer: new trends on solution methodologies and applications. J Heat Transf 134(3). doi:10.1115/1.4005131.
10.1115/1.4005131...
). Since these types of IHCPs formulations require intensive iterative procedures, many efforts of the scientific community are focused on reducing the computational costs related to techniques, algorithms and mathematical formulations (Orlande et al. 2014ORLANDE HRB, DULIKRAVICH GS, NEUMAYER M, WATZENIG D & COLAÇO MJ. 2014. Accelerated Bayesian inference for the estimation of spatially varying heat flux in a heat conduction problem. Num Heat Transf Part A Appl 65(1): 1–25. doi:10.1080/10407782.2013.812008.
10.1080/10407782.2013.812008...
, Vakili & Gadala 2009VAKILI S & GADALA MS. 2009. Effectiveness and efficiency of particle swarm optimization technique in inverse heat conduction analysis. Num Heat Transf Part B Fund 56(2): 119–141. doi:10.1080/10407790903116469.
10.1080/10407790903116469...
). Parallel to these efforts, the use of computational intelligence emerged as a possible approach for solving IHCPs, such as Genetic Algorithms (Garcia et al. 1998GARCIA S, GUYNN J & SCOTT EP. 1998. Use of genetic algorithms in thermal property estimation: part II-simultaneous estimation of thermal properties. Num Heat Transf Part A Appl 33(2): 149–168. doi:10.1080/10407789808913932.
10.1080/10407789808913932...
, Imani et al. 2006IMANI A, RANJBAR A & ESMKHANI M. 2006. Simultaneous estimation of temperature-dependent thermal conductivity and heat capacity based on modified genetic algorithm. Inv Probl Sci Eng 14(7): 767–783. doi:10.1080/17415970600844242.
10.1080/17415970600844242...
) and Artificial Neural Networks (ANN) (Balaji & Padhi 2010BALAJI C & PADHI T. 2010. A new ANN driven MCMC method for multi-parameter estimation in two-dimensional conduction with heat generation. Int J Heat Mass Transf 53(23-24): 5440–5455. doi:10.1016/j.ijheatmasstransfer.2010.05.064.
10.1016/j.ijheatmasstransfer.2010.05.064...
, Czél et al. 2014CZÉL B, WOODBURY KA & GRÓF G. 2014. Simultaneous estimation of temperature-dependent volumetric heat capacity and thermal conductivity functions via neural networks. Int J Heat Mass Transf 68: 1–13. doi:10.1016/j.ijheatmasstransfer.2013.09.010.
10.1016/j.ijheatmasstransfer.2013.09.010...
, Shiguemori et al. 2004SHIGUEMORI EH, DA SILVA JDS & CAMPOS VELHO HF DE. 2004. Estimation of initial condition in heat conduction by neural network. Inv Probl Sci Eng 12(3): 317–328. doi:10.1080/10682760310001598599.
10.1080/10682760310001598599...
, Soeiro et al. 2004SOEIRO F, SOARES PO, CAMPOS VELHO HF & SILVA NETO AJ. 2004. Using neural networks to obtain initial estimates for the solution of inverse heat transfer problems. In: Inverse Problems, Design an Optimization Symposium, p. 358-363. ), which are considered soft computing approaches, as opposed to classical ones.

The use of ANNs has been permeating the most eclectic types of applications, such as engineering, financial analysis, image and speech recognition, medicine, among many others. This success can be credited to several factors, one of them being the notable improvements and developments of different ANN architectures, such as Autoencoders (AE), Recurrent Neural Networks (RNN), Convolutional Neural Networks (CNN), Extreme Learning Machines (ELM), and so on. The use of any of these architectures must take into consideration its specific features along with characteristics of the problem being solved.

Recently, several of these architectures emerged as viable techniques for defect characterization using IHCPs with infrared thermography. Hu et al. ( 2019HU C, DUAN Y, LIU S, YAN Y, TAO N, OSMAN A, IBARRA-CASTANEDO C, SFARRA S, CHEN D & ZHANG C.. 2019. LSTM-RNN-based defect classification in honeycomb structures using infrared thermography. Infr Phys Technol. Infrared Physics & Technology 102: 103032. doi:10.1016/j.infrared.2019.103032.
10.1016/j.infrared.2019.103032...
) employed a long short term memory RNN for automatically classifying different types of common defects occurring in honeycomb materials. Later, Fang & Maldague ( 2020FANG Q & MALDAGUE X. 2020. A method of defect depth estimation for simulated infrared thermography data with deep learning. Appl Sci 10(19): 6819. doi:10.3390/app10196819.
10.3390/app10196819...
) used a Gated Recurrent Units (GRU) for depth prediction in composite material samples. Kaur et al. ( 2021KAUR K, MULAVEESALA R & MISHRA P. 2021. Constrained Autoencoder based Pulse Compressed Thermal Wave Imaging for Sub-surface Defect Detection. IEEE Sens J doi:10.1109/JSEN.2021.3056394.
10.1109/JSEN.2021.3056394...
) introduced a novel Constrained Autoencoder in order to reduce the dimensionality of temporal thermographic images for defect depth estimation in steel plates. Marani et al. ( 2021MARANI R, PALUMBO D, GALIETTI U & D’ORAZIO T. 2021. Deep learning for defect characterization in composite laminates inspected by step-heating thermography. Opt Laser Eng 145: 106679. doi:10.1016/j.optlaseng.2021.106679.
10.1016/j.optlaseng.2021.106679...
) used step-heating thermography combined with CNN in order to describe defects in composite laminates. Xu & Hu ( 2021XU L & HU J. 2021. A Method of Defect Depth Recognition in Active Infrared Thermography Based on GRU Networks. Appl Sci 11(14): 6387. doi:10.3390/app11146387.
10.3390/app11146387...
) performed defect depth identification by using a method based also on GRUs.

Besides the cited innovation in networks architectures, the most widely used in the field of heat transfer engineering is the simple feed-forward back propagation Multilayer Perceptron (MLP). Indeed, this network has been proved to be a great general purpose approximator in countless applications, but it lacks some advantages that a specific deep neural network, such as CNNs, AEs, RNN and others may give.

In this context, AEs (or Replicator networks) are great tools designed for reducing the dimension of any given data (Haykin 2009HAYKIN SS. 2009. Neural networks and learning machines. 3rd ed. Upper Saddle River, NJ: Pearson Education , Hinton & Salakhutdinov 2006HINTON GE & SALAKHUTDINOV RR. 2006. Reducing the dimensionality of data with neural networks. Science 313(5786): 504–507. doi:10.1126/science.1127647.
10.1126/science.1127647...
). Such network have the input equal to the output, but it main aspect is a bottleneck-like hidden layer (latent space) which is smaller in dimension than the input-output. When properly trained, the original data can be encoded into this smaller dimension latent space, then the information in this space can be decoded back into the original information size. Furthermore, AEs can be specially designed as a noise reduction network, also known as Denoising Autoencoders (DAE). This can be performed by giving an noisy data as input, passing it through the latent space, then outputting a noiseless version of that same data. If well trained, the information stored in the latent space is an almost-noiseless representation of the original data.

In this work the IHCP of estimating the spatially varying thermal properties of multilayered media is solved via an ANN, where the input data are temperature measurements of a exposed surface and the output is a set of discrete points representing the thermal properties of different materials. More specifically, an encode-decode architecture is used, where the normalized experimental temperature and the thermal properties are encoded separately with respectives AE, then the latent space of each one of them are used as input and output, respectively, for the multi layer perceptron (MLP) neural network. Although this approach imposes the training of three ANN (two AEs and one MLP), which yields an extra computational cost when compared to a simple MLP, the use of the AE shows advantages such as the robustness to different architectures, as presented in the obtained results.

In order to test this approach, the problem of contact failures identification was chosen, where two materials are joined with an adhesive layer. Gaps of air can form in this layer, meaning that the adhesion process is not perfect (Abreu et al. 2014ABREU LAS, ORLANDE HRB, KAIPIO J, KOLEHMAINEN V, COTTA RM & QUARESMA JNN. 2014. Identification of contact failures in multilayered composites with the Markov chain Monte Carlo method. J Heat Transf 136(10). doi:10.1115/1.4027364.
10.1115/1.4027364...
, 2016ABREU LAS, COLAÇO MJ, ORLANDE HRB & ALVES CJS. 2016. Thermography detection of contact failures in double layered materials using the reciprocity functional approach. Appl Therm Eng 100: 1173–1178. doi:10.1016/j.applthermaleng.2016.02.078.
10.1016/j.applthermaleng.2016.02.078...
, Mascouto et al. 2020MASCOUTO FS, KNUPP DC, ABREU LAS, SANCHES EL & SPHAIER LA. 2020. Detection of contact failures employing combination of integral transforms with single-domain formulation, finite differences, and Bayesian inference. Num Heat Transf Part A Appl 77(6): 599–618. doi:10.1080/10407782.2020.1713637.
10.1080/10407782.2020.1713637...
). Identifying the thermal properties of this layer containing gaps is crucial to access the adhesion quality in manufacturing and preventive maintenance.

In this work, the training data set is generated by solving the direct problem, i.e., the temperature profiles are simulated and obtained for different configurations of failures. Furthermore, the temperature data containing simulated noise is used to train a DAE, where its output is the noise free version of the same data. Due to its dimensionality and noise reduction, this approach can greatly benefit the use of increasingly larger data, as presented in the results. Also, some effort goes into fine tuning the parameters of the ANN itself, such as activation functions, layer sizes, number of neurons, and so on. As mentioned, this is where the encoded-decoded approach shows its promising characteristics, by yielding robust results for different types of configurations.

Direct Problem Formulation and Solution

Consider the Lx×Ly multilayered rectangular medium presented in Fig. 1 composed by three different layers. Such representation can model the bonding of two materials, where Layer 1 and 3 are joined by an adhesive layer shown as Layer 2. In the configuration of Fig. 1, two defects are illustrated in the adhesive layer. Both left and right boundaries are considered adiabatic, the constant heat flux q is prescribed at the bottom boundary, and the top one is considered exposed, with ambient temperature of T and natural convection heat transfer coefficient h. Temperature measurements Texp are acquired at the exposed boundary, i.e., y=Ly.

Figure 1
Schematic representation of the multilayered medium.

Using a single-domain formulation (Mascouto et al. 2020MASCOUTO FS, KNUPP DC, ABREU LAS, SANCHES EL & SPHAIER LA. 2020. Detection of contact failures employing combination of integral transforms with single-domain formulation, finite differences, and Bayesian inference. Num Heat Transf Part A Appl 77(6): 599–618. doi:10.1080/10407782.2020.1713637.
10.1080/10407782.2020.1713637...
) and taking in consideration the mentioned assumptions, the heat conduction problem can be written as

x ( k ( x , y ) T ( x , y , t ) x ) + y ( k ( x , y ) T ( x , y , t ) x ) = ρ ( x , y ) c p ( x , y ) T ( x , y , t ) t (1a)
T ( x , y , t ) x | x = 0 = 0 , T ( x , y , t ) x | x = L x = 0 (1b,c)
k ( x , y ) T ( x , y , t ) y | y = 0 = q , k ( x , y ) T ( x , y , t ) y | y = L y + h T ( x , L y , t ) = h T (1d,e)
T ( x , y , 0 ) = T (1f)

where h is the convection heat transfer coefficient at the top boundary, and q is the prescribed heat flux at the bottom boundary. The materials properties k(x,y),ρ(x,y), and cp(x,y) are the thermal conductivity, specific mass, and specific heat, respectively. These spatially dependent properties are modeled in such way to represent the different materials of the multilayered medium. Furthermore, the initial temperature is considered equal to the ambient temperature T.

If all the thermophysical properties and all the conditions and contact failures geometries are known, the problem described by Eq. (1) can be solved yielding the temperature profile T(x,y,t). In this work this solution is obtained by using the built-in “NDSolve" routine of the Wolfram Mathematica 11.0 software. This routine is executed with the Finite Element Method option and the mesh is generated with rectangular user defined uniform cell size. This routine returns an interpolated temperature profile T(x,y,t) for the whole Lx×Ly medium and for times ranging from t=0 to t=tmax, which is a prescribed maximum time instant.

Inverse Problem Formulation

Consider that the materials of the multilayered medium presented in Fig. 1 are known, but the existence and position of possible defects in Layer 2 are unknown. In this case, the inverse problem is formulated based on the assumption that Layer 2 is composed entirely by adhesive or by adhesive with air gaps (defects). This leads to the definition of an auxiliary function γ(x) which represents the thermal properties k(x,y) and ρ(x,y)cp(x,y) of Layer 2, defined as

γ(x)=ϕ(x)ϕgapϕlayerϕgap(2)
where ϕ(x) is the related thermophysical property, that is, the thermal conductivity k(x,y) or the volumetric heat capacity ρ(x,y)cp(x,y), of Layer 2 materials. Assuming that there is no variation in the y direction within the materials of Layer 2, γ(x) is considered only as an x variable function. Also, the subscripts “gap" and “layer" represent the related material thermal properties. The function γ is used algorithmically as a way to translate the thermal properties to the spatial dependent boolean function “gap, no-gap", “0" or “1", respectively. Furthermore, γ can be reversely translated to the thermal properties k(x,y) and ρ(x,y)cp(x,y) of Layer 2, since the values ϕgap and ϕlayer are known.

In order to obtain estimates of the gaps positioning in Layer 2, an ANN is used to represent the inverse model, i.e. the network receives temperature experimental data as input and outputs a vector representing discrete points of the auxiliary function γ(x). Experimental Data. [paragraph:simulated_data] In order to simulate real measurements of temperature, the experimental data is computationally generated, where random numbers r drawn from a normal distribution centered at zero with known standard deviation σexp are added to solution of the direct problem, i.e.,

T i , j = T ( x i , L y , t j ) with i = 1 , 2 , . . . , N , j = 1 , 2 , . . . , M (3a)
T i , j exp = T i , j + r , r N ( 0 , σ exp 2 ) , with i = 1 , 2 , . . . , N , j = 1 , 2 , . . . , M (3b)

where xi represents the positioning of temperature acquisition along the x coordinate with the total number of N points, and tj the time instants with total number of M levels. Therefore, the total number of experimental data considered is Nexp=N×M.

Inverse Problem Solution with Artificial Neural Networks

As previously mentioned, the IHTP of this work has the objective of estimating the positioning of gaps that can occur in Layer 2 of the medium (see Fig. 1). This task is performed by estimating discrete points of the auxiliary function γ(x): this is the ANN output. The input consists of transient temperature measurements that are encoded into a latent space of an AE, which is used with the purpose of reducing the dimension and noise on the experimental data. This latent space serves as the input to a fully connected MLP network that yields another latent space of the same size of the first one. This last latent space is then decoded to a full representation of the auxiliary function γ. With this approach, both latent spaces are used as input and output, respectively, to the fully connected MLP neural network.

In the following subsections, the two AEs and MLP construction are explained and detailed. Firstly, the temperature and γ AE are presented. Next, the complete network architecture is showcased. Lastly, general aspects necessary to build the network and the simulated data set are shown.

Autoencoders

Temperature Autoencoder (T-AE). Consider the AE structure presented in Fig. (2). The input A is fully connected to one hidden layer, represented by HA, which is the latent space. Then HA is fully connected to the output A^. These connections are all modified by weights, biases, and activation functions.

Figure 2
Schematic representation of the Temperature Denoising Autoencoder.

As presented in Fig. (2), the AE input A and output A^ are normalized versions of Texp and T, respectively (see Eq. 3). The components of A are the normalized version of Texp, and the ones for A^ are the normalized version of T, both ranging between 1 and 1. These linear normalizations are performed with limits of Tmin and Tmax, which are the minimum and maximum exact temperature values of the entire training set (see subsection entitled General Aspects of the Artificial Neural Network for details).

The T-AE implemented in this work has only one hidden layer, which is the latent space itself, but in some cases, AE’s can have many hidden layers as the problem requires. Using the architecture with the single hidden layer, tests with different simulated experimental data led to good approximations, making the increased network with more layers unnecessary.

Choosing the latent space HA dimension is crucial to a good performance of the AE. If too small, some information will be lost, leading to a poor representation of the temperature data. On the other hand, if its dimension is too large, training requires more examples and the problem can become less efficient overall. Therefore a comparison of different HA dimensions is performed in this work. Adhesive Layer Properties Autoencoder (γ-AE). The adhesive layer thermophysical properties is represented by the auxiliary function γ(x), which ranges from 0 to 1. The profile of this function represents different material properties and its behavior indicates where there are gaps in the layer.

As presented in Fig. 3, this second AE is used to represent γ and has input G, output G^ and latent space HG. The input G is composed by discrete points used to represent γ across the x direction. G^ is the exact representation of G, therefore, they are trained to be equal. The latent space HG dimension is chosen to be equivalent as HA, i.e., equal to the T-AE latent space dimension.

Figure 3
Schematic representation of the Adhesive Layer Properties Autoencoder using the Auxiliary Function γ(x).

It is important to note that this Adhesive Layer Properties AE has to encode less information than the T-AE presented, since it has only to output 0 or 1, so tests with dimensions smaller than the T-AE size yielded good results, which is expected. Nevertheless, in this work the choice of maintaining both latent space sizes equal is kept.

Final Artificial Neural Network Model

In order to transform HA into HG, a fully connected MLP is used, i.e., this network model has to translate latent spaces information from temperature measurements to the adhesive layer properties in terms of γ(x). Therefore, HA and HG are the input and output to the MLP, respectively, where hidden layers are connected and the information passing is modified by weight, biases and activation functions.

The final ANN architecture is displayed in Fig. (4), where it is possible to see the normalized experimental temperature input A. This input is encoded into HA, which is then passed and transformed into HG to finally be decoded as G^. With this approach the actual γ(x) estimation process utilizes only machine information, which is an advantage to obtain more robust models. Also, it is important to note that such network model allows to an increased dimension of the temperature input data, which is desirable with a transient problem. To illustrate this aspect of the ANN model, different time instants are considered and compared in this work.

The training of the Fully Connected MLP in Fig. (4) is performed by using the same training set that trained all the other models. To perform this task, firstly, the two AE’s must have been properly trained. Then, all the experimental temperature of the training set is encoded into HA and all the respective discretized γ(x) profiles are encoded into HG. With this, the pairs HAHG are assembled and the model can be trained.

Figure 4
Schematic representation of the complete network model.

General Aspects of the Artificial Neural Network

To build the training data set for the three ANN, a list of input-output pairs is necessary. One training pair for the T-AE is composed by the rule A A^, which are the normalized versions of the experimental data Texp and the exact temperature data T, respectively. Similarly, the material property AE uses the pair G G, which is the discrete representation of the exact γ used to generate that particular T and Texp. After completing the training of both AE’s, the set for the MLP can be build with input HA and output HG by using the two AE’s to translate the original training set into the related latent space.

To build this training data set, the direct problem is solved with a given random thermophysical properties profile for Layer 2 containing gaps, which is represented as the auxiliary function γ. Furthermore, it is important to follow some criteria that represents the physics of the problem in order to avoid unnecessary computations (Czél et al. 2014CZÉL B, WOODBURY KA & GRÓF G. 2014. Simultaneous estimation of temperature-dependent volumetric heat capacity and thermal conductivity functions via neural networks. Int J Heat Mass Transf 68: 1–13. doi:10.1016/j.ijheatmasstransfer.2013.09.010.
10.1016/j.ijheatmasstransfer.2013.09.010...
). In this work, the training data set is built by assuming that:

  • There is a maximum number of gaps in Layer 2, i.e. Ngap=1,2,...,Mgap;

  • There is a minimum gap size, namely Lmin;

  • The gaps only occur in the region of xminxxmax, where xmin and xmax are chosen to avoid gaps being too close to the boundaries.

  • The T-AE and final MLP dataset are composed of multiple copies of the original set, with each one containing different level of noise for Texp.

These rules are a way to generate meaningful information to the ANN model. For example, if the gap is too narrow, it will not influence the temperature on the exposed surface of the medium, specially considering experimental noise. Furthermore, the last rule is included in order to generalize the noise reduction aspect of the T-AE, by giving different noise levels in the data set.

Results and discussion

Direct Problem Solution

Considering a steel plate for Layer 1, adhesive for Layer 2, composite protection for Layer 3 and air as the inclusions in the possible defects (gaps), the respective thermophysical properties used are displayed in Table I.

Table I
Thermal Properties for the Multilayered Medium.

Furthermore, to obtain the results of the direct problem, the prescribed heat flux, the natural convection coefficient, and the ambient temperature are defined as q=10000 W/m2, h=15.0 W/m2K, and T=20.0 °C, respectively. The medium dimensions are defined as Lx=0.2 m and Ly=0.071 m, and Layer 2 is vertically positioned in the range 0.05y0.051 m.

Training, Validation and Testing Sets Generation

All the training and validation data are generated with 1, 2 or 3 gaps in Layer 2, i.e. Mgap=3. For each particular number of gaps, 2000 randomly positioned gaps configurations are used to obtain the respective transient temperature profile, yielding a total of 6000 direct problem evaluations. The training and validation data are then split into a 90% to 10% relation, respectively. Therefore, from this total set, 5400 are used for training and 600 for validation.

To build the simulated experimental data used for the training process of the T-AE and the MLP, four levels of random noise are considered for σexp, they are 0.05 °C, 0.1 °C, 0.15 °C and 0.20 °C (see Eq. 3b). This type of measurement is typically obtained via infrared thermography techniques (Knupp et al. 2013KNUPP DC, NAVEIRA-COTTA CP, ORLANDE HRB & COTTA RM. 2013. Experimental identification of thermophysical properties in heterogeneous materials with integral transformation of temperature measurements from infrared thermography. Exper Heat Transf 26(1): 1–25. doi:10.1080/08916152.2011.631079.
10.1080/08916152.2011.631079...
). Hence, four sets of 6000 input-output pairs compose the final training data, yielding 24000 examples (21600 for training and 2400 for validation). It is important to notice that the γ-AE training is performed with the original 6000 auxiliary function profiles.

The testing data is generated similarly to the training and validation set. It is composed of 150 random positioned defects (50 for 1, 2, and 3 simultaneous failures). The same noise levels are considered, therefore the original 150 data is copied yielding 600 examples for the T-AE and MLP testing - the γ AE keeps the original 150.

In order to illustrate how efficiently the AE can store information in its latent space, the results of this work compares three problems with different time instants for the measured experimental data. For M=3 time instants, 600 s, 800 s, and 1000 s are considered, for M=5, 600 s, 800 s, 1000 s, 1200 s and 1400 s, and lastly, for M=7, the instants range from 600 s to 1800 s in steps of 200 s. For every instant, N=321 points in the x coordinate are considered. So, for problems with 3, 5 and 7 time levels, the ANN must deal with a input of 963, 1605 and 2247 data, respectively. These time instants and total amount of input data are displayed in Table II.

Table II
Time Instants and number of temperature data used.

It is important to notice that the model proposed in this work considers increasingly time levels of physical experiment. In other words, for M=3, M=5 and M=7, the final time of temperature acquisition is 1000 s, 1400 s and 1800 s, respectively. Therefore there is a trade off between amount of data and temporal information available. At the lower spectrum, for M=3, less temperature readings are used, making the model training and configuration more efficient. On the other hand, for M=7, the larger data set can contain useful transient information of the problem, in exchange of training and model configuration efficiency. Hence, M=5 can be considered as the middle point of the two mentioned.

The exact maximum temperature values necessary to normalize the training data set between 1 and 1 are different for the sets with 3, 5 and 7 time instants, they are: 35.5718 °C, 47.8281 °C and 60.5670 °C, respectively. The minimum temperature obtained is equal for the three sets, with value of 21.1653 °C, which occurs for t=600 s.

Temperature Autoencoder Results

The T-AE implementation is performed using the Keras library, from Python 3.9 programming language. The Logistic Sigmoid activation function is selected for the hidden layer and the output layer is kept as linear. The Adam optimizer is selected as implemented within the Keras library and the Logarithmic Hyperbolic Cosine (log-cosh) is chosen as the loss function. Such function has the advantage of being totally differentiable, with a L2 norm behavior for small values of its argument, and as L1 for larger values (Queiroz et al. 2021QUEIROZ RM, RIBEIRO EG, SANTOS ROSA N & GRASSI JR V. 2021. On deep learning techniques to boost monocular depth estimation for autonomous navigation. Robot Auton Sys 136: 103701. doi:10.1016/j.robot.2020.103701.
10.1016/j.robot.2020.103701...
).

Early Stopping is selected as the stopping criterion. When the minimum loss function value found during training for the validation set does not change for 200 epochs, the iterative procedure is stopped. Then, the weights and biases obtained for the best epoch are used. This approach avoids over fitting of the training data (Chollet 2017CHOLLET F. 2017. Deep Learning with Python. New York, NY: Manning Publications. ). Furthermore, the maximum number of epochs is set to 6400, but none of the configurations tested reached that value. Finally, the batch size selected is 256, which is a relatively large size, but results have shown to be a good compromise between quality and speed of training.

In Table III the results obtained for the T-AE with M=3 time instants are displayed for different latent spaces dimensions, namely “AE dim" column. The table displays the Validation and Testing Mean Squared Errors obtained (MSE), along with the number of Training Epochs needed for that particular result. The result displayed as “Max. error" represents the maximum relative temperature error. This error is obtained with the following expression

Max. error=max(|T̂i,j,kTi,j,kTi,j,k|)×100(4)
where T̂ and T are the denormalized T-AE output and exact temperatures, respectively, both obtained for each testing data k, time level j and point in space i.

Table III
Temperature Autoencoder (T-AE) MSE results for 3 time instants and different Autoencoder (AE) dimensions.
Table IV
T-AE MSE results for 5 time instants and different AE dimensions.

As it is possible to see in Table III, the AE dimensions such as 20, 40 and 60 generated the worst MSE results for the validation and testing set. Nevertheless, it is possible to notice that, with the exception of the AE with 20 dimensions, every configuration yielded a maximum error lower than 1%. Here is also possible to see that, in general, the epochs necessary to reach the stopping criterion ranges from 1021 to 1875, which can be considered a relatively low discrepancy. From the Test MSE criterion, the best configuration was the 80-dim AE. Here it is important to note that the MSE metric tends to increase larger errors, due to its squared argument. This is a desirable feature for comparing different ANN configurations, the best networks tend to have small and low discrepancy among them, whereas the worst configurations will generate larger values with higher discrepancy among themselves.

The T-AE results for M=5 and M=7 time instants are presented in Table IV and V, respectively. For both results it is possible to see that the validation and testing MSE tends to decrease as the AE dimension increases. For M=5, the maximum error values are also lower than 1% with exception of the 20-dim AE, which is not true for M=7, since only the dimensions 120, 140 and 160 generated lower than 1% maximum relative errors. This can be explained in the light of the amount of data considered for 7 time instants and their range, i.e., since all the data is normalized between -1 and 1, the more information one fits inside this range, the more difficult becomes to the network train with efficiency. Nevertheless, the lowest validation and testing MSE generated the lowest maximum relative error for M=7 (120-dim: 4.08, 3.97 and 0.63%).

Table V
T-AE MSE results for 7 time instants and different AE dimensions.

MLP Results

To construct the complete ANN architecture, it is necessary to specify the elements for the MLP and γ-AE. Firstly, the activation functions for γ-AE latent space and output layer are the ReLU and Logistic Sigmoid functions, respectively. Since γ-AE outputs the auxiliary function, which ranges between 0 and 1, the Logistic Sigmoid is a reasonable choice that acts similar to a filter, limiting the output values inside that range and, furthermore, results with a linear output yielded undesired oscillations around the exact value. Regarding the loss function for training, the Binary Cross Entropy was selected, which have a good behavior for optimizing values that can be either 0 or 1. Secondly, the MLP is composed of four hidden layers, with the Hyperbolic Tangent (tanh) activation function for each and linear output layer. It is important to note that the MLP input and output sizes are equal as the latent spaces of both AE’s, as previously mentioned. The output layer activation is linear and the loss is calculated using the log-cosh functional. Other activation functions such as ReLu, Leaky ReLU and Logistic Sigmoid were tested for the hidden layers, but the best result (including faster training) were obtained with the tanh function. The stopping criterion for training for both γ-AE and MLP is similar as the T-AE, after 200 epochs with no improvements on the validation set loss function ADAM optimizer iterative process stops.

Tables VI, VII and VIII present the MSE between the exact discretized γ profile and the respective ANN final output for different MLP sizes (number of neurons per hidden layer). Therefore these “Test MSE" results compares the decoded γ-AE output obtained from the MLP output, which MSE error is displayed as “Test HG".

In Table VI the results for M=3 time instants are displayed, which were obtained with the 80-dim T-AE (therefore the γ-AE is also used with 80 dimensions). Similarly, Tables VII and VIII show the results for 5 and 7 time instants, with T-AE of 160-dim and 120-dim, respectively.

In Table VI, for M=3, the MLP size that yielded the lowest γ MSE was the one with 200 neurons per hidden layer. Nevertheless, the 80 MLP size, which is equal as the used T-AE and γ-AE for this problem, generated the second best result. As it is possible to see, the epochs used for training decreases as the MLP size increases, which is indication that overfitting was avoided. Another important aspect to observe is that the best HG MSE was not observed for the same MLP size as the best γ MSE, which indicates that the γ-AE accepts small differences in the latent spaces elements in order to decode it into the same output.

For M=5 and M=7 in Tables VII and VIII, the best found MLP size was the one with 40 neurons per hidden layer. Similarly to the results in Table VI, the training epochs needed also decreases when the MLP size increases, and the best HG MSE was not observed for the best γ MSE. It is worth noting the the respective best γ output MSE for 3, 5 and 7 time instants were 0.01396935, 0.01423047 and 0.01425704, respectively, indicating that the problem considering 3 instants generated the lowest error. Nevertheless, the three results are relatively similar.

Table VI
MSE for M=3, 80-dim AE, and different MLP sizes.
Table VII
MSE for M=5, 160-dim AE, and different MLP sizes.
Table VIII
MSE for M=7, 120-dim AE, and different MLP sizes.

Detection of Contact Failures

The auxiliary function γ estimation using the best ANN obtained in Tables VI, VII, and VIII are presented in this subsection. To illustrate how the model can generalize the input data with different noise levels, the results are considered for σexp of 0.1 °C and 0.2 °C.

In order to illustrate the T-AE noise reduction capability, the comparison among simulated experimental temperature data, the denormalized T-AE output, and the exact temperature is displayed in Fig. 5a, 6a, 7a, 8a, 9a, and 10a for different contact failures configurations. The simulated experimental temperature is generated with noise level of σexp=0.2 °C and M=3 (three time instants). The T-AE output is obtained with the 20-dim latent space configuration. In such figures, the γ(x) presented indicates the exact auxiliary function profile used to obtain the temperature displayed above each one. For the sake of visualization, the results with T-AE’s of 5 and 7 time instants are not displayed.

Although the noise level of σexp=0.2 °C was the largest used in training, the approximation between the exact temperature profiles and the denormalized T-AE output is excellent, as it is possible to see for every configuration with one (Fig. 5a and 6a), two (Fig. 7a and 8a), and three (Fig. 9a and 10a) contact failures. This indicates that the AE has good dimensionality reduction qualities and it can, indeed, be used as a feature extractor tool, which stores the most important information necessary to reconstruct the temperature data without noise in the latent space.

Figure 5
T-AE output (a) and γ(x) estimation (b) for one small contact failure configuration.
Figure 6
T-AE output (a) and γ(x) estimation (b) for one large contact failure configuration.
Figure 7
T-AE output (a) and γ(x) estimation (b) for two small contact failures configuration.
Figure 8
T-AE output (a) and γ(x) estimation (b) for one small and one large contact failures configuration.
Figure 9
T-AE output (a) and γ(x) estimation (b) for three small contact failures configuration.
Figure 10
T-AE output (a) and γ(x) estimation (b) for one large and two small contact failures configuration.

Figure 5b and 6b presents the results for 1 single contact failure, being one small and one large, respectively. Both results shows a good approximation of the estimated γ profile with the exact one. The small gap in Fig. 5b has 10 mm, which should be the hardest to detect due to lack of sensitivity, and shows very reasonable fit for all time instants considered. Therefore, this configuration with only one gap can be considered the easiest to estimate.

Now considering two defects in Layer 2, the results are presented in Fig. 7b and 8b for two small gaps and for one small and one large gap, respectively. Results in Fig. 7b shows two small gaps of 10 mm length and separated by another 10 mm. This configuration show very good results for M=5 and M=7 time instants, for both noise level considered. Although the results for M=3 are good, some undesired oscillation in the second gap is found. For one small and one large gaps, the results presented in 8b shows very good agreement to the exact values, considering that they are more far apart from each other. The large gap has 30 mm in length and the small 10 mm.

Finally, the results considering three gaps in Fig. 9b and 10b. The former is composed of three small gaps, and the latter with two small and one large. These configuration are the most difficult to estimate, due to the increase non-linearity of the problem: it makes the AE’s work harder to extract more information from the temperature data. All the gaps in Fig. 9b have 10 mm of length and are separated by 20 mm intervals. Here the agreements of results with exact values are not exactly centered, but, nevertheless, the three gaps positioning is well identified. Specially considering the result with M=5 and 7, which generated less oscillations. For two small and one large gaps, the results showed some difficult in estimation. The best agreement is obtained for M=3 for both noise levels, as indicated by results in the subsection entitled MLP Results. In this configuration, the model for M=5 presented some oscillations. Results with M=7 found the correct positioning of the two small gaps, but not the right size of them by estimating larger defects than the exact ones.

The computational times necessary for this investigation can be divided into three categories: training data generation, training of the models, and activation of the neural network. The system used to perform all executions is equipped with a AMD Ryzen™ 5 4600H processor, GeForce® GTX 1650 graphics processor, 8GB of RAM, and Microsoft Windows 10 operational system. Some aspects of computational times for each of the mentioned categories are the following:

  • The training data generation is the most intensive computational task of this work. The time necessary to evaluate sequentially the 6000 different direct problem solutions, as described in the subsection General Aspects of the Artificial Neural Network, was 15.6 hours. Although this task was performed sequentially, it has a fixed number of iterations and can be easily parallelized in order to reduce this computational time.

  • The computational times spent to train the neural networks are varied, given the different architectures and networks types. For the sake of illustration, the training of MLP’s with 160 and 200 neurons per layer of M=7 time instants took 3 minutes and 22 seconds, and 2 minutes and 57 seconds, respectively, with both using 628 and 558 epochs to reach the stopping criterion - see Table VIII.

  • The computational cost of performing the final estimation of γ(x) with the three networks is the main motivation for using the approach proposed. For instance, the estimation for M=7 and σexp=0.1 °C, as presented in Fig. 6, takes an average of 0.113 seconds.

Conclusions

The present work used an ANN in order to estimate the positioning of defects that can occur in the bonding between two materials. To reduce the dimension of the 1D transient thermography experimental data, an AE was used. Different configurations of AE were tested considering data from 3, 5 and 7 time instants, which corresponds to 963, 1605 and 2247 temperature readings, respectively. The main aspects of the proposed approach for solving the defect positioning estimation problem may be summarized as the following items:

  • Speed of estimation. When compared to classical approaches such as Maximum Likelihood or Bayesian Methods, the use of ANN’s for estimating parameters of the physical problem has a great advantage in computational efficiency. Usually, such classical methods have cumbersome iterative processes that are not practical in some real scenarios. On the other hand, the estimation with ANN’s can instantly be obtained.

  • Reduced dimensions of temperature data. Since the problem solved in this work is transient, several time instants must be included as the ANN input. This work deals with this problem by encoding such transient temperature profiles in order to facilitate the work of estimating failures. Thus, the reduction in dimension was necessary to extract important features of the noisy data and work with a manageable ANN.

  • Increased estimation resolution. Since an AE is also used to reduce the dimension of the output, a larger amount of points can be used to represent the varying thermophysical properties of the adhesive layer.

  • Fixed computational cost for training data generation. Each training data is generated by solving the direct problem once. Therefore, the amount of examples needed to train the model is generated only once. For every result presented in this work, the same 6000 direct problem solutions were used.

Moreover, further investigations must continue in two main subdivisions: Validation of the proposed ANN model with real experimental data, and construction of a ANN model that considers the complete 2D transient thermography image as experimental data.

ACKNOWLEDGMENTS

The authors acknowledge the financial support provided by Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES, Finance Code 001), Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq) and Fundação Carlos Chagas Filho de Amparo à Pesquisa do Estado do Rio de Janeiro (FAPERJ). LCSJ acknowledges CAPES also for the scholarship for his stay in the University of Granada (Grant CAPES/PrInt No. 88887.469279/2019-00). This work has been also partially funded by the projects PID2020-112754 GB-I00 (Spanish Ministry of Economy and Competitiveness and funds from the European Regional Developement Fund, ERDF) and B-TIC-640-UGR20 (Regional Govern of Andalusia, Spain).

  • ABREU LAS, COLAÇO MJ, ORLANDE HRB & ALVES CJS. 2016. Thermography detection of contact failures in double layered materials using the reciprocity functional approach. Appl Therm Eng 100: 1173–1178. doi:10.1016/j.applthermaleng.2016.02.078
  • ABREU LAS, ORLANDE HRB, KAIPIO J, KOLEHMAINEN V, COTTA RM & QUARESMA JNN. 2014. Identification of contact failures in multilayered composites with the Markov chain Monte Carlo method. J Heat Transf 136(10). doi:10.1115/1.4027364
  • BALAJI C & PADHI T. 2010. A new ANN driven MCMC method for multi-parameter estimation in two-dimensional conduction with heat generation. Int J Heat Mass Transf 53(23-24): 5440–5455. doi:10.1016/j.ijheatmasstransfer.2010.05.064
  • BECK JV, BLACKWELL B & CLAIR JR CRS. 1985. Inverse heat conduction: Ill-posed problems. John Wiley & Sons.
  • CHOLLET F. 2017. Deep Learning with Python. New York, NY: Manning Publications.
  • CZÉL B, WOODBURY KA & GRÓF G. 2014. Simultaneous estimation of temperature-dependent volumetric heat capacity and thermal conductivity functions via neural networks. Int J Heat Mass Transf 68: 1–13. doi:10.1016/j.ijheatmasstransfer.2013.09.010
  • DANES F, GARNIER B & DUPUIS T. 2003. Predicting, measuring, and tailoring the transverse thermal conductivity of composites from polymer matrix and metal filler. Int J Thermophys 24(3): 771–784. doi:10.1023/A:1024096401779
  • FANG Q & MALDAGUE X. 2020. A method of defect depth estimation for simulated infrared thermography data with deep learning. Appl Sci 10(19): 6819. doi:10.3390/app10196819
  • GARCIA S, GUYNN J & SCOTT EP. 1998. Use of genetic algorithms in thermal property estimation: part II-simultaneous estimation of thermal properties. Num Heat Transf Part A Appl 33(2): 149–168. doi:10.1080/10407789808913932
  • HAYKIN SS. 2009. Neural networks and learning machines. 3rd ed. Upper Saddle River, NJ: Pearson Education
  • HINTON GE & SALAKHUTDINOV RR. 2006. Reducing the dimensionality of data with neural networks. Science 313(5786): 504–507. doi:10.1126/science.1127647
  • HU C, DUAN Y, LIU S, YAN Y, TAO N, OSMAN A, IBARRA-CASTANEDO C, SFARRA S, CHEN D & ZHANG C.. 2019. LSTM-RNN-based defect classification in honeycomb structures using infrared thermography. Infr Phys Technol. Infrared Physics & Technology 102: 103032. doi:10.1016/j.infrared.2019.103032
  • IMANI A, RANJBAR A & ESMKHANI M. 2006. Simultaneous estimation of temperature-dependent thermal conductivity and heat capacity based on modified genetic algorithm. Inv Probl Sci Eng 14(7): 767–783. doi:10.1080/17415970600844242
  • JAYASEELAN DD, XIN Y, VANDEPERRE L, BROWN P & LEE WE. 2015. Development of multi-layered thermal protection system (TPS) for aerospace applications. Comp Part B Eng 79: 392–405. doi:10.1016/j.compositesb.2015.04.017
  • KAIPIO JP & FOX C. 2011. The Bayesian framework for inverse problems in heat transfer. Heat Transf Eng 32(9): 718–753. doi:10.1080/01457632.2011.525137
  • KAUR K, MULAVEESALA R & MISHRA P. 2021. Constrained Autoencoder based Pulse Compressed Thermal Wave Imaging for Sub-surface Defect Detection. IEEE Sens J doi:10.1109/JSEN.2021.3056394
  • KNUPP DC, NAVEIRA-COTTA CP, ORLANDE HRB & COTTA RM. 2013. Experimental identification of thermophysical properties in heterogeneous materials with integral transformation of temperature measurements from infrared thermography. Exper Heat Transf 26(1): 1–25. doi:10.1080/08916152.2011.631079
  • MARANI R, PALUMBO D, GALIETTI U & D’ORAZIO T. 2021. Deep learning for defect characterization in composite laminates inspected by step-heating thermography. Opt Laser Eng 145: 106679. doi:10.1016/j.optlaseng.2021.106679
  • MASCOUTO FS, KNUPP DC, ABREU LAS, SANCHES EL & SPHAIER LA. 2020. Detection of contact failures employing combination of integral transforms with single-domain formulation, finite differences, and Bayesian inference. Num Heat Transf Part A Appl 77(6): 599–618. doi:10.1080/10407782.2020.1713637
  • MOURA NETO FD & SILVA NETO AJ. 2013. An introduction to inverse problems with applications. Springer Science & Business Media. doi:10.1007/978-3-642-32557-1
  • ORLANDE HRB. 2012. Inverse problems in heat transfer: new trends on solution methodologies and applications. J Heat Transf 134(3). doi:10.1115/1.4005131
  • ORLANDE HRB, DULIKRAVICH GS, NEUMAYER M, WATZENIG D & COLAÇO MJ. 2014. Accelerated Bayesian inference for the estimation of spatially varying heat flux in a heat conduction problem. Num Heat Transf Part A Appl 65(1): 1–25. doi:10.1080/10407782.2013.812008
  • QUEIROZ RM, RIBEIRO EG, SANTOS ROSA N & GRASSI JR V. 2021. On deep learning techniques to boost monocular depth estimation for autonomous navigation. Robot Auton Sys 136: 103701. doi:10.1016/j.robot.2020.103701
  • SARKAR D, HAJI-SHEIKH A & JAIN A. 2015. Temperature distribution in multi-layer skin tissue in presence of a tumor. Int J Heat Mass Transf 91: 602–610. doi:10.1016/j.ijheatmasstransfer.2015.07.089
  • SHIGUEMORI EH, DA SILVA JDS & CAMPOS VELHO HF DE. 2004. Estimation of initial condition in heat conduction by neural network. Inv Probl Sci Eng 12(3): 317–328. doi:10.1080/10682760310001598599
  • SOEIRO F, SOARES PO, CAMPOS VELHO HF & SILVA NETO AJ. 2004. Using neural networks to obtain initial estimates for the solution of inverse heat transfer problems. In: Inverse Problems, Design an Optimization Symposium, p. 358-363.
  • SU J, CERQUEIRA DR & WANG S-Z. 2012. Thermal design of multi-layered composite pipelines for deep water oil and gas production. Int J Comp Appl Technol 43(3): 248–259. doi:10.1504/IJCAT.2012.046311
  • VAKILI S & GADALA MS. 2009. Effectiveness and efficiency of particle swarm optimization technique in inverse heat conduction analysis. Num Heat Transf Part B Fund 56(2): 119–141. doi:10.1080/10407790903116469
  • XU L & HU J. 2021. A Method of Defect Depth Recognition in Active Infrared Thermography Based on GRU Networks. Appl Sci 11(14): 6387. doi:10.3390/app11146387

Publication Dates

  • Publication in this collection
    01 Aug 2022
  • Date of issue
    2022

History

  • Received
    13 Dec 2021
  • Accepted
    17 Apr 2022
Academia Brasileira de Ciências Rua Anfilófio de Carvalho, 29, 3º andar, 20030-060 Rio de Janeiro RJ Brasil, Tel: +55 21 3907-8100 - Rio de Janeiro - RJ - Brazil
E-mail: aabc@abc.org.br