This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Digital twin, physics-based model, and machine learning applied to damage detection in structures

T.G. Ritto and F.A. Rochinha
[email protected], [email protected]
Dept. of Mechanical Engineering
Universidade Federal do Rio de Janeiro
Brazil

Abstract

This work is interested in digital twins, and the development of a simplified framework for them, in the context of dynamical systems. Digital twin is an ingenious concept that helps on organizing different areas of expertise aiming at supporting engineering decisions related to a specific asset; it articulates computational models, sensors, learning, real time analysis, diagnosis, prognosis, and so on. In this framework, and to leverage its capacity, we explore the integration of physics-based models with machine learning. A digital twin is constructed for a damaged structure, where a discrete physics-based computational model is employed to investigate several damage scenarios. A machine learning classifier, that serves as the digital twin, is trained with data taken from a stochastic computational model. This strategy allows the use of an interpretable model (physics-based) to build a fast digital twin (machine learning) that will be connected to the physical twin to support real time engineering decisions. Different classifiers (quadratic discriminant, support vector machines, etc) are tested, and different model parameters (number of sensors, level of noise, damage intensity, uncertainty, operational parameters, etc) are considered to construct datasets for the training. The accuracy of the digital twin depends on the scenario analyzed. Through the chosen application, we are able to emphasize each step of a digital twin construction, including the possibility of integrating physics-based models with machine learning. The different scenarios explored yield conclusions that might be helpful for a large range of applications.

Keynotes: digital twin, physical based model, machine learning classifier, damage identification, structural dynamics.

1 Introduction

The fourth industrial revolution is related to data, artificial intelligence, robotics, internet of things, and much more [1]. Every day there are more sensors, data, and computer power at our disposal. In such a context, the interest in developing digital twins for engineering systems is rapidly growing. More specifically, here we would like to treat structural dynamic problems, such as wind turbine dynamics and drill string dynamics, operating in locations of difficult access and harsh environments.

There is no consolidated view on what digital twins are [2, 3, 4]. However, it is consensual that a digital twin is a virtual representation of a specific physical asset, which uses data collected from this asset to connect digital and physical parts. The potential benefits within such a generic framework have sparked much attention from different industrial segments. In section 2 of the present paper, a digital twin conceptual framework and its ingredients are detailed.

A virtual representation means a physics-based computational model and/or a machine learning model. The former is the usual model obtained from the combination of first principles (e.g. Newton’s second law) with closure models (e.g. constitutive model), and the later is obtained through fitting data to some mathematical structure; for instance, neural networks [5], or any other data-driven model [6, 7, 8].

Recently, lots of researches are dedicated to leverage computational models, by integrating physics-based models with machine learning [9, and references therein]. The present paper aims at tackling this integration in the context of building digital twins, and some strategies are depicted in section 3. It should be noticed that the use of a physics-based model ensures physical interpretability (very desired for engineering systems), and machine learning models are very adapted to data, and suited to real time application [10].

A simple structure, modeled as a lumped-parameter system, is the asset to be tracked by a digital twin in the present analysis. The steps of constructing a digital twin are emphasized, and a strategy to combine physics-based models with machine learning is employed. The main ingredients of our digital twin are: (i) computational model, (ii) uncertainty quantification, (iii) calibration/update of the model using data from the physical twin.

The contributions of the present paper are twofold: (i) to construct a digital twin conceptual framework, in the context of structural dynamics, to detect damage in structures, and (ii) to assess a strategy to combine physics-based models with machine learning, for the system under analysis. In this attempt, section 2 gives a general view of digital twins, and section 3 summarizes ideas related to physics-based models and machine learning. There is an effort of the research community to explore these possibilities, and their many variations, in several applications.

The strategy employed here is detailed in section 4. It considers a stochastic physics-based computational model (with different damage scenarios) to train a machine learning classifier (our digital twin). This methodology might be applied whenever a computational model is useful to understand (and catalog) different behaviors (or failures) of a dynamical system. The digital twin classifier will be connected to the physical twin, supporting real time engineering decisions.

There is an increasing body of literature in line with the main topics under analysis here. Bigoni and Hesthaven [11] employ a strategy similar to the one analyzed in the present work. They called it simulation-based anomaly detection, and propose it in the context of structural health monitoring; where Support Vector Machines are used to detect anomalies. Zhao et al [12] apply deep learning (auto-encoder, convolution networks, etc.) in machine health monitoring systems. A deep learning based framework for prognostics and health monitoring is also proposed by Booyse et al [13]. Karvea et al [14] propose a digital twin approach for damage diagnosis, damage prognosis, and optimization for crack growth. Finally, Kapteyn et al [15] develop a digital twin for an aerial vehicle, and consider a damaged structure. The computational physics-based model is updated using a machine learning classification technique called optimal trees [16]. After training the classifier with datasets from a library that contains models with different levels of degradation (damage), the machine learning classifier is able to choose the best candidate model, given new experimental data. A state-of-art digital twins for modelling and simulation in engineering dynamics applications can be found in [4].

This paper starts establishing, in section 2, the digital twin framework for our analysis. Then, alternatives to combine physics-based models with machine learning, within the context of a digital twin, are detailed in section 3. The physical twin for damage detection is developed in section 4, and it is assessed in section 5. Finally, the concluding remarks are made in the last section of the paper.

2 A Digital twin framework

Jones et al. [2] performed a systematic literature review on Digital Twin, and, after consolidation, reached thirteen main characteristics. Some of them are exposed here, adapted to our specific context. Wagg et al. [4] presents a state-of-art more related to the present work, modeling and simulation in engineering dynamics applications. Some elements exposed in [4] will be explored further in this section.

Figure 1 shows a schematic representation of a digital twin conceptual framework. Measurements are taken from the physical twin (wind turbine) to calibrate/update the digital twin. The digital twin is composed of a computational model (physics-based and/or machine learning models) and a stochastic layer to take into account uncertainties. Note that the inverse problem (calibration) is stochastic, which can be tackled with the Bayesian methodology [17], for instance.

A digital twin is well suited to real time applications [10], and it can be useful on system automation [18]. Its goal is to support engineering and operational decisions, and it can be used for: (i) feeding information to actuators and operational parameters, (ii) setting up alarms, (iii) performing diagnosis, (iv) performing prognosis (what if questions and estimate residual life), and also (v) designing a system with improved performance, (vi) optimizing sensors/actuators locations, etc.

A digital twin is more than a computational model. It is a computational model, that takes into account uncertainties, and is calibrated with data measured from its physical twin (a specific asset). It is updated whenever needed to track the life cycle of the asset; i.e., it evolves with time, replicating the history of performance and degradation of its twin. Some pretty well established disciplines are part of a digital twin framework; for example, finite element model [19], neural networks [5], digital signal processing [20], structural health monitoring [21], uncertainty quantification [22], inverse problem [23], model updating [24], and verification and validation [25]. All of these areas should be coordinated inside the digital twin framework to provide results that support decisions concerning a specific asset goals.

Refer to caption
Figure 1: Digital twin framework. Measurements are taken from the physical twin (wind turbine) to calibrate/update the digital twin. The digital twin is composed of a computational model (CM), which integrates physics-based and machine learning (ML) models. And, a stochastic layer is added to the deterministic model, to take into account uncertainties.

In this digital twin conceptual framework, we consider three mandatory ingredients: (i) computational model, (ii) uncertainty quantification, and (iii) calibration using laboratory (off-line) and field online data from the asset (physical twin). Some remarks should be made:

  • There are physical-to-virtual and virtual-to-physical connections. Data collected from the physical twin are used to calibrate/update the digital twin. And, the predictions of the digital twin are used to support decision concerning the physical twin operation (parameter values, control strategy, degradation evaluation, etc).

  • The digital twin must represent the main characteristics of the physical twin within a given confidence.

  • The level of confidence of the digital twin predictions should be calculated by means of a stochastic analysis (e.g. probability of failure). Note: there are other approaches to model uncertainties (e.g. fuzzy sets [26]), but they are not considered here.

  • The computational model might be physics-based and/or machine learning. Specifically, we are interested in leveraging the computational model integrating both types of modeling; see section 3.

  • Even if based on physical principles, issues related to parameters and model-form (epistemic) uncertainties should be taken care of [22, 27].

The framework presented in this section can be expanded to capture a wider picture. For instance (details can be found in [4, and references therein]): combination of multiple models (different physics, fidelity, scales) of the structure, laboratory tests using components from the structure, multiple teams of experts, computations of the subsystems in parallel, commissioning, process control, software integration and management, technical and organisational problems, autonomous manufacturing. The digital twin can start in the design phase of its physical counterpart, and pass through the manufacturing process. Wagg et al [4] proposed a W model for product design, where a specific virtual prototyping stage is included and used as the basis for the digital twin.

3 Integrating physics-based models with machine learning

Let us start this section commenting on some general characteristics of physics-based and machine learning models. Before, it should be mentioned that, while constructing a model, and to extract most of it, the role of the specialist (expert knowledge) is extreme valuable. The model hypotheses and goals must be clearly stated, and the characteristics of the target engineering application will determine the type and sophistication of modeling. One should ask questions, such as: How complex is the system? Is the physics well understood? What kind of data are available, and what is the acquisition rate? Is it a real time application? What is the level of uncertainty (of inputs, parameters, models, and outputs)? Etc.

A data-driven analysis is paramount in the digital twin realm. Both physics-based and machine learning models must be calibrated/trained with experimental or field data. Part of the data should be separated for calibration/training, and the remainder for validation/testing. More complex models have more parameters, what makes the calibration procedure more difficult. Depending on the problem, normalization of data and parameters, and regularization of objective functions, are highly recommended [7].

Some characteristics of a physics-based model are listed below:

  • It is constructed using first principles (e.g. Newton’s second law) combined with phenomenological closure models (e.g. constitutive models, friction models, damping models, boundary conditions, joints)

  • Each parameter has a clear physical interpretation

  • One can build high fidelity time consuming computational models of complex engineering systems [28]

  • One can calibrate the model at a given operational condition and use it to analyze different scenarios (good extrapolation capacity)

In summary, we can build high fidelity and interpretable physics-based models, that can be used to analyze a multiplicity of new scenarios. Nevertheless, uncertainty is an important issue for complex systems, with joints, friction, and multi-physics, for example. And if part of the physics is not known, another mathematical structure should take place. Finally, they can have high computational cost, which is critical for real time applications. To alleviate this last issue, reduced-order models serving as proxies for the original expensive one might be constructed [29, 30, 31].

Some characteristics of a machine learning model are listed below [7]:

  • It is obtained from some mathematical architecture that is not based on physical laws; it is a linear or a complicated nonlinear transformation of inputs into outputs

  • Usually its parameters are seldom physically interpretable

  • It can fit well experimental data; for instance, a neural network with a single hidden layer can approximate any continuous function, with arbitrary accuracy [7, 5]

  • Over-fitting is an issue and should be avoided

  • The capacity of extrapolation is limited; if the machine learning model is not trained for a specific scenario, it can not generally make predictions about it

In summary, we can build pretty fancy data-driven machine learning models, that can be used to tackle complex engineering problems. No physics laws are needed, and, if enough data is available, their predictions can be accurate. Furthermore, they run fast, amenable to real time applications. However, they are often limited to the training domain, and it is usually hard to extrapolate the results with confidence. In addition, they can not be used to design a new system. Finally, they usually need more data for the training (calibration) to compensate the absence of an initial mathematical structure (materialized as a physical law); and finding an optimal network architecture, for example, remains an art.

It seems to make a lot of sense to integrate physics-based models with machine learning to take advantage of their pros and diminish their cons. And that is exactly what a lot of researchers are doing [9, and references therein], setting the realm of Physics aware Machine Learning. There are many possibilities for this integration, some of which are out of the scope of the present work, such as using machine learning to discover governing equations [32], using machine learning to solve partial differential equations [33], incorporating physical constraints into the machine learning [34, 35], transfer learning using physics-based model to pre-train the machine learning [36], embedding physical principles into the neural network design [37], analyzing the mismatch between a physics based model and data [34].

We are particularly interested in using physics-based models for intensively simulating different damage scenarios to be tracked during the online operation of the digital twin. The resulting significant amount of data is employed to train a machine learning [15, 11, 38] that serves as a digital twin to detect damage, or to select an appropriate model, in the structural system. It is important to highlight the role played by a physics-based model in amplifying the interpretability of machine learning tools, what is crucial for the success of a digital twin in achieving the main goals of the detection system: locating and evaluating the severity of the damage. Another critical point in such a strategy for putting together an effective digital twin relies on conciliating the production of accurate information through the simulations and the required amount of data to train properly the machine learning classifier. That can be achieved by a judicious design of the computer model, as we illustrate in our example of a digital twin prototype to be introduced in the next section.

4 A Digital Twin to damage detection in structures

In the present section, we give form to the above ideas and concepts that delineates a digital twin through a prototype dedicated to damage detection. Identification of damage in structures is conveniently phrased as an inversion problem, in which the input data is obtained by sensors and the outputs are damage localization and severity. Models are used to connect inputs and outputs. The resulting mathematical problem is usually cumbersome to be solved in real time applications. Here, to cope with the need of providing online responses, we follow a nonstandard approach. We train a machine learning classifier using synthetic data produced by a physics-based model that replicates, at a certain extent, the real structure dynamic response. The role of such classifier is to relate the inputs (sensors signals) to plausible damage scenarios obtained though the model in an off-line stage.

In addition to the profile of digital twin described above, we have to consider the three main goals of damage detection: identify the presence, location and severity of damaged regions. Our proposal addresses those issues with the help of simulations employing a physics-based model. Indeed, [11] also uses synthetic data for training their classifier but differs in the sense that localization is achieved by assigning to each critical area one digital twin and positioning sensors accordingly. Moreover, one can also extend such a diagnosis perspective, by estimating the residual life of the structure, pinpointing risks to its future functioning. This prognosis phase is not addressed here.

A bar structure is used to illustrate our prototype of physical twin. Then, a discrete computational model is built to represent the physical twin, and to support the interpretability of the machine learning tools embedded in it. The physics-based model with a localized damage is used to simulate different degraded scenarios of the structure. For this purpose, a stochastic model is developed to construct datasets. We go step-by-step in the design of the prototype in the sequence of this section.

4.1 The Physical twin

Refer to caption
Figure 2: Sketch of the physical twin.

The physical twin considered here consists of a slender dynamic structure that vibrates axially, as depicted in the form of a sketch in Fig. 2. Indeed, in the present application, we do not have a real physical twin. Hence, a prismatic bar governed by the following partial differential equation, considered here as a high-fidelity model of the asset, is used to emulate it. Such equation results from the combination of the balance of linear momentum with a elastic constitutive equation.

ρA2u(x,t)t2EA2u(x,t)x2=f(x,t),\rho A\frac{\partial^{2}u(x,t)}{\partial t^{2}}-EA\frac{\partial^{2}u(x,t)}{\partial x^{2}}=f(x,t)\,, (1)

where ρ\rho is the density of the material, EE is the elasticity modulus, AA is the cross section area, uu is the displacement field that is a function of the independent variables xx (position) and tt (time), and ff if the force per unit length. This partial differential equation is discretized by means of the finite element method, from which we obtained the discretized system (linear shape functions are employed)

[Mpt]𝐮¨pt(t)+[Cpt]𝐮˙pt(t)+[Kpt]𝐮pt(t)=𝐟pt(t),[M_{pt}]\ddot{\mathbf{u}}_{pt}(t)+[C_{pt}]\dot{\mathbf{u}}_{pt}(t)+[K_{pt}]\mathbf{u}_{pt}(t)=\mathbf{f}_{pt}(t)\,, (2)

with the appropriate initial conditions, and where the subscript ptpt stands for physical twin. The usual mass, stiffness and proportional damping matrix ([C]=α0[M]+β0[K])[C]=\alpha_{0}[M]+\beta_{0}[K])) are shown in the equation; in which α0\alpha_{0} and β0\beta_{0} are positive constants. A fixed-free boundary condition is considered, and the system is rewritten in the frequency domain

𝐮^pt(ω)=(ω2[Mpt]+jω[Cpt]+[Kpt])1𝐟^pt(ω),\hat{\mathbf{u}}_{pt}(\omega)=(\omega^{2}[M_{pt}]+j\omega[C_{pt}]+[K_{pt}])^{-1}\hat{\mathbf{f}}_{pt}(\omega)\,, (3)

where j=1j=\sqrt{-1}. The system is discretized in forty finite elements (elementary matrices are in Appendix A), and A=4×104m3A=4\times 10^{-4}m^{3}, L=1mL=1m, ρ=7850kg/m3\rho=7850kg/m^{3}, E=210×109PaE=210\times 10^{9}Pa, α0=1×103\alpha_{0}=1\times 10^{3}, and β0=1×107\beta_{0}=1\times 10^{7}. The first three natural frequencies, computed from the mass and stiffness matrices (generalized eigenvalue problem) are {1293,3881,6476}\{1293,3881,6476\} Hz and the first damping ratios are {6.3,2.4,1.8}\{6.3,2.4,1.8\}%.

4.2 The discrete computational model

Refer to caption
Figure 3: Physics-based computational model.

The computational model is physics-based, and is constructed using a 6-DOF lumped parameter description, as displayed in (Fig. 3):

𝐮^cm(ω)=(ω2[Mcm]+jω[Ccm]+[Kcm])1𝐟^cm(ω),\hat{\mathbf{u}}_{cm}(\omega)=(\omega^{2}[M_{cm}]+j\omega[C_{cm}]+[K_{cm}])^{-1}\hat{\mathbf{f}}_{cm}(\omega)\,, (4)

where the mass and the stiffness matrices, [Mcm][M_{cm}] and [Kcm][K_{cm}], are described in Appendix B, and [Ccm]=α0[Mcm]+β0[Kcm][C_{cm}]=\alpha_{0}[M_{cm}]+\beta_{0}[K_{cm}]. The first three natural frequencies, considering m=0.3925kgm=0.3925kg and k=4914×105N/mk=4914\times 10^{5}N/m, are {1358,3999,6398}\{1358,3999,6398\} Hz and the first damping ratios are {6.0,2.4,1.9}\{6.0,2.4,1.9\}%.

A lumped parameter model fulfills the requirements raised before. It is able to furnish a significant amount of training data with reasonable computational costs, capturing the essential dynamic response of the physical counterpart, within the frequency range of interest, as will be made clear in the sequence. Moreover, it also equipped to handle the two main issues of damage detection as each spring can be associate with the average stiffness of segments in the continuum model. Indeed, that can be easily established by using in the finite element discretization a lumped mass distribution, linear elements, and a coarse grid. Such connectivity between continuum and discrete models helps on improving the interpretability of the machine learning classifier.

Figure 4 shows the frequency response of the physical twin (Eq. 3) and the computational model (Eq. 4), where a decorrelated Gaussian noise was added to the physical twin, with zero mean and standard deviation of σ=5×104\sigma=5\times 10^{-4} m, to reproduce usual monitoring conditions. A force with magnitude of 10×10310\times 10^{3} N is applied at the right end of the system, and the response is obtained at the same position, which corresponds to the sixth degree-of-freedom of the discrete computational model. The response of this computational model does not match the one of the physical twin, but it is yet quite representative. The amplitudes of the responses are comparable and the error in the first natural frequencies are lower than 5%.

Refer to caption
Figure 4: Frequency responses of physical twin and the deterministic computational model.

4.3 The Stochastic computational model

To augment the predictive capability of the computational model by quantifying its confidence level, we take into account uncertainties originated from different sources endowing the original model with a stochastic structure, assuming probabilistic parametric characterizations. We admit that AA, EE, ρ\rho, and LL are independent Uniform random variables 𝒰(min,max)\sim\mathcal{U}(min,max).

Therefore, mass, stiffness, and damping matrices are now random 𝐌cm\mathbf{M}_{cm}, 𝐊cm\mathbf{K}_{cm}, 𝐂cm\mathbf{C}_{cm}, as is the response of the system 𝐔^cm\hat{\mathbf{U}}_{cm}, denoted in bold. The stochastic model is, thus, expressed in the form of its FRF:

𝐔^cm(ω)=(ω2[𝐌cm]+jω[𝐂cm]+[𝐊cm])1𝐟^cm(ω),\hat{\mathbf{U}}_{cm}(\omega)=(\omega^{2}[\mathbf{M}_{cm}]+j\omega[\mathbf{C}_{cm}]+[\mathbf{K}_{cm}])^{-1}\hat{\mathbf{f}}_{cm}(\omega)\,, (5)

Figure 5 shows the physical twin response together with the 95% statistical envelope of the stochastic computational model, where the bounds of each random Uniform variable were taken to plus and minus five percent of the nominal value (e.g. [0.95A,1.05A][0.95A,1.05A]). The response of the computational model is now able to get closer to the physical twin response. Of course, the price we pay is to deal with a statistical envelope; which represents the level of confidence.

Refer to caption
Figure 5: Frequency responses of the physical twin and the stochastic computational physics-based model.

More sophisticated techniques to model uncertainties and to calibrate a stochastic model can be found in [23, 22]. In the present investigation, we will focus on the digital twin general methodology.

4.4 Damage parametrization

We assume that damage manifests, at the macroscopic scale, as a local loss of stiffness [39], and a simple damage model is considered for the analyses [40]. In the context of our computer model, damage is parametrized by means of a scalar β[0,1]\beta\in[0,1] that modulates the reduction in the spring stiffness. The stiffness value is multiplied by β\beta at the damage location (corresponding spring), where d=1βd=1-\beta, with dd the percentage of the damage expressing its severity. For instance, if a 10% damage is considered at the second spring, k2=βk=(10.10)kk_{2}=\beta k=(1-0.10)k (Eq. 6); and the stiffness matrix is recalculated, as well as the damping matrix. Moreover, we have the constraint β˙0\dot{\beta}\geq 0 to be consistent with the underlying physics of damage process [41]. In the present application, we do not follow damage evolution, therefore there is no need to take this constraint into account.

[K]d2=[k+βkβk0000βkβk+kk0000k2kk0000k2kk0000k2kk0000kk].{}_{d_{2}}[K]=\left[\begin{array}[]{cccccc}k+\beta k&-\beta k&0&0&0&0\\ -\beta k&\beta k+k&-k&0&0&0\\ 0&-k&2k&-k&0&0\\ 0&0&-k&2k&-k&0\\ 0&0&0&-k&2k&-k\\ 0&0&0&0&-k&k\\ \end{array}\right]\,. (6)

A healthy structure is characterized by d=0d=0 in all springs (in all continuum segments), and hence its FRF provides a reference to identify damaged structures in the frequency band of interest. Figure 6 shows the response of the deterministic system without damage together with five damage situations: damage of 20% at the (1) first spring, (2) second spring, (3) third spring, (4) fourth spring, and (5) fifth spring (no damage was considered in the last spring). The response changes depending on the degree of freedom observed, and on the frequency of the applied force. For instance, at 7000 Hz, the response corresponding to the healthy structure (black) touches the curve with 20% of damage at the fourth spring (blue). But, at the same frequency, the response of the sixth DOF shows black and blue curves far apart; see Fig. 6 (b). Such initial studies justify the choice of vibration amplitudes, due to their sensitivity to damage, as principal features to be employed in the identification technique.

Refer to caption

(a) Refer to caption(b)

Figure 6: Frequency responses of the deterministic computational model, corresponding to different damage scenarios. (a) response at DOF #1, (b) response at DOF #6.

4.5 Constructing the training dataset

A supervised learning is employed to train the classifier, where the dataset XdataX_{data} is constructed using many samples of the displacements of our discrete system: U^1\hat{U}_{1}, U^2\hat{U}_{2},…, as needed. Other responses could be used, such as velocity, acceleration, or deformation, where the most appropriate type of response will depend on the application and on the sensors at disposal. The dataset XdataX_{data} (features) and the associated damaged scenario ylabely_{label} (labels) are the input training pair for the machine learning classifier, a central piece of our digital twin, as will be detailed in the sequence.

The dataset structure comprises nn random samples for each one of the mm damage scenarios, which are generated using our stochastic computational and damage models (Eqs. 5 and 6). The response of each sample (i)(i) is organized in a row of matrix XdataX_{data}. For example, the damage scenario d1d_{1} can be 10% of damage at spring # 1. Hence,

Xdatad1=[U^1(1)U^2(1)U^1(n)U^2(n)],{}_{d_{1}}X_{data}=\left[\begin{array}[]{cccccc}\vspace{.2cm}\hat{U}_{1}^{(1)}&\hat{U}_{2}^{(1)}&...\\ \vspace{.2cm}\hfil&...&\\ \vspace{.2cm}\hat{U}_{1}^{(n)}&\hat{U}_{2}^{(n)}&...\\ \end{array}\right]\,, (7)

which is associated to the label d1d_{1}. The multivariate nature of the digital twin allows, partially and implicitly, to address two main issues of the damage detection: localization and severity. The final dataset XdataX_{data} is a (mn×l)(mn\times l) dimension matrix, which is a composition of the nn random samples of each mm damage scenarios (including the healthy structure), with ll displacements measured:

Xdata=[Xdatad1Xdatad2Xdatadm].X_{data}=\left[\begin{array}[]{cccccc}\vspace{.2cm}{}_{d_{1}}X_{data}\\ \vspace{.2cm}{}_{d_{2}}X_{data}\\ \vspace{.2cm}...\\ \vspace{.2cm}{}_{d_{m}}X_{data}\\ \end{array}\right]\,. (8)

As done in section 4.3, the parameters AA, EE, ρ\rho, and LL, are modeled as independent Uniform random variables. In addition, we will allow fluctuations of the damage variable, the forcing frequency, and the measured displacements, to allow a more generalized digital twin; because the damage of the physical twin might be a little higher or a little lower than the nominal value, the excitation frequency might be a little different, and there might be noisy measurements. So, the damage dd is also modeled as a Uniform random variable with bounds [0.95d,1.05d][0.95d,1.05d], the excitation frequency ωf\omega_{f} idem [0.95ωf,1.05ωf][0.95\omega_{f},1.05\omega_{f}], and a zero mean Gaussian random noise is added to the amplitude of the response, σ=10×107\sigma=10\times 10^{-7}.

4.6 Construction of the digital twin - machine learning classifier

As mentioned in the last section, the stochastic computational model is used to simulate the response of the system with different damage locations and corresponding intensities. From these simulations, a dataset is constructed to train the digital twin, which is a machine learning classifier, see Fig. 7. The trained classifier can be used in the digital twin platform to deploy a diagnosis and quickly warn the operation if there is damage, and where it is located.

Refer to caption
Figure 7: The physics-based computational model is used to construct a dataset that is used to train a machine learning classifier.
Classifier Accuracy
Quadratic Discriminant 93.3%
SVM (quadratic) 93.1%
SVM (linear) 92.2%
SVM (cubic) 86.0%
Linear Discriminant 84.8%
KNN 81.8%
SVM (Gaussian) 80.6%
Ensemble (Bagged Trees) 77.9%
Decision Tree 61.2%
Ensemble (RUSBoosted Trees) 39.5%
Table 1: Accuracy of different classifiers.

There are several machine learning classifiers, such as k-nearest neighbor, linear/quadratic discriminants, support vector machines, decision tree, and ensemble methods [42]. We tested a number of well established classifiers, and Table 1 shows their accuracy111The parameters of some classifiers can be tuned to achieve a better performance, but this is not the goal of the present work. The idea is to show that the Quadratic Discriminant has a good performance, and other classifiers, such as SVM, perform similarly. Note: quadratic discriminant takes 0.06s to run, while SVM (quadratic) takes 11.97s.,222For the decision Tree, the maximum number of splits is 100, and the split criterion is the Gini’s diversity index. For the Support Vector Machine (SVM), the Box constraint level is 1, and the multi-class method is one-vs-one. For the Gaussian kernel, the scale is 0.61. For the K nearest neighbor (KNN), the number of neighbors is one, and the distance metric is Euclidean. For the Bagged trees, the number of learners is 30. For the RUSBoosted trees, the maximum number of slips is 20 and the learning rate is 0.1., evaluated by the percentage of correctly classified inputs amongst the whole set of data, and where a 5-fold cross validation was employed [42]. The frequency of excitation used was ωf=3800\omega_{f}=3800 Hz, and the damage dd=20%. This set of parameters yield good results. In the next sections, we analyze the impact of changing some parameters, on the accuracy of the machine learning model.

The classifier that best performed, observing this accuracy criterion and for the current application, was the quadratic discriminant. Therefore, we present a high-level overview of the quadratic discriminant analysis (QDA), that also naturally accommodates our probabilistic approach. A classifier 𝒢\mathcal{G} aims to provide a partition of the multidimensional input domain associate to KK classes k\mathcal{M}_{k}, to be defined by the user, accordingly with previously established objectives.

In QDA, the Bayes formula is used to compute the posterior probability of the kk-th class given the input vector 𝐱\mathbf{x}:

P(=k|X=𝐱)=fk(𝐱)πkl=1Kfl(𝐱)πl,P(\mathcal{M}=k|X=\mathbf{x})=\frac{f_{k}(\mathbf{x})\pi_{k}}{\sum_{l=1}^{K}f_{l}(\mathbf{x})\pi_{l}}\,, (9)

where πk\pi_{k} is the prior probability of the kk-th class (number of times this class is observed divided by the total number of observations) with k=1Kπk=1\sum_{k=1}^{K}\pi_{k}=1, and fkf_{k} is the probability density function of the kk-th class, which is assumed to be Gaussian:

fk(x)=1(2π)n/2|Σk|1/2exp{12(xμk)TΣk1(xμk)},f_{k}(x)=\frac{1}{(2\pi)^{n/2}|\Sigma_{k}|^{1/2}}\exp{\{-\frac{1}{2}(x-\mu_{k})^{T}\Sigma_{k}^{-1}(x-\mu_{k})\}}\,, (10)

in which the mean μk\mu_{k} and the covariance matrix Σk\Sigma_{k} of class kk are estimated from the dataset. Note that there is no extra parameter to tune. To judge if a point xx is from class kk or ll, the probability ratio is verified, P(G=k|X=x)/P(G=l|X=x)P(G=k|X=x)/P(G=l|X=x). If this ratio is greater than one, it is classified as class kk, i.e., the class with higher probability is chosen. If the covariance matrices of each class are not the same (which is the present case), then the discriminant function is quadratic, hence the name of the method.

To summarize, Fig. 8 shows the digital twin framework for our application. In this framework, physics-based and machine learning models are integrated as follows. Measurements are taken from the physical twin (bar structure) to calibrate a stochastic computational model (physics-based), which is used to train the digital twin (a machine learning classifier). The physics-based model has the advantage of been interpretable, and able to emulate different scenarios of damage. And the machine learning model has the advantage of running fast. The trained digital twin classifier could be used for a real time application. That is, it would be connected to the physical twin, processing new dynamic signals, and, whenever a signal is associated with a damage, there would be a warning.

Refer to caption
Figure 8: Digital twin framework. Measurements are taken from the physical twin (bar structure) to calibrate a stochastic computational model, which is used to train the digital twin (a machine learning classifier). The digital twin classifier is used for real time operations. Whenever a signal is associated to a damage, there is a warning.

5 Assessing the Digital Twin

To understand many aspects of the proposed digital twin, several analysis are performed in this section.

5.1 Dataset 1: 20% of damage at each spring

Figure 9 shows the dataset used to train the digital twin, considering a damage of 20% at each spring. We do not analyse scenarios in which two (or more) springs are damaged simultaneously. Two hundred samples were collected for each one of the six configurations analyzed: healthy and damage placed at each i-th spring (i=1,..,5). To improve the training conditions, the data is normalized by subtracting the mean and dividing by the standard deviation of all samples.

Refer to caption

(a) Refer to caption(b) Refer to caption(c) Refer to caption(d)

Figure 9: Training data generated with the stochastic digital twin with excitation frequency of 3800 Hz. Responses at DOFs (a) #1 and #6, (b) #2 and #4, (c) #3 and #5, (d) #5 and #6.

The plots in Fig. 9 display some separation within the data, and, thus, this observed clustering bears potential to an input domain partition by a classifier. Particularly, Fig.9(a), which presents the correlation between displacements of first and sixth DOFs, shows red dots (20% of damage at the first spring) at the bottom of the cloud of points. And, Fig. 9(b), which presents the displacements of the second and fourth DOFs, shows green dots (20% of damage at the third spring) at the top of the cloud of points. It is not clear, a priori, if the machine learning model will be able to classify correctly the damage location. Amazingly, the accuracy of the quadratic discriminant strategy is very high: 93.3%.

One hundred samples, not used in the training, were employed to validate the digital twin. The confusion matrix [43] is presented in Table 2. Each row represents the real damage scenario, and the columns display the results of the classifier. The first row shows that 86 out of the 100 (row sum) samples for the healthy structure are correctly classified. One sample is misclassified as d2d_{2} (damage at the second spring), 10 incorrectly classified as d3d_{3}, and so on. Moving to the third row, the same statistics for damage detection can be obtained for class d2d_{2}: 97 out of the 100 samples are correctly classified, 1 sample is incorrectly classified as healthy, and 2 samples are incorrectly classified as d3d_{3}. Therefore, the diagonal terms are related to correctly classified samples, and the off-diagonal terms are related incorrectly classified samples. There is a lot of information in this matrix, such as the quantification of true positives and false negatives. The accuracy is excellent when the damage is at the first spring (100%), at the second spring (97%), at the fourth spring (99%) and at the fifth spring (100%). The most difficult damage to locate is at the third spring (78%). Other information that can be taken from this matrix is the probability of considering a damage when the system is healthy (false-positive), which is (1-86%)=14%. And the probability of indicating a healthy structure, when the third spring is damaged (false-negative), is 19%. Overall, the results are quite satisfactory.

healthy d1d_{1} d2d_{2} d3d_{3} d4d_{4} d5d_{5}
healthy 86 0 1 10 3 0
d1d_{1} 0 100 0 0 0 1
d2d_{2} 1 0 97 2 0 0
d3d_{3} 19 0 0 78 2 1
d4d_{4} 0 0 0 0 99 1
d5d_{5} 0 0 0 0 0 100
Table 2: Confusion matrix considering the healthy structure, and 20% of damage at each spring.

One might argue that 20% is too much damage. The level of damage will surely affect the performance of the digital twin, as it will be confirmed in the sequence. Actually, the performance of the digital twin depends on several factors, such as combinations of:

  • damage severity

  • damage location

  • number of sensors (and sensors arrangement)

  • excitation location

  • excitation frequency

  • signal to noise ratio

  • level of parameter (and model) uncertainties

The next sections will analyze the performance of our digital twin varying these factors.

5.2 Dataset 2: 10% of damage at each spring

If a damage of 10% is considered, the accuracy of the digital twin gets lower (80.3%), which is expected since, as damage decreases, it is harder to distinguish among the different scenarios. At the limit, when d=0, all scenarios converge to d0d_{0} (healthy structure).

The confusion matrix is given in Table 3, showing a lower performance, but the main conclusions are similar: the accuracy is very good when the damage is at the first spring (95%), at the fourth spring (90%) and at the fifth spring (90%). The most difficult damage to locate is at the third spring (80%). The probability of considering a damage when the system is healthy (false-positive) is (1-64%)=36%. And the probability of indicating a healthy structure, when the third spring is damaged (false-negative), is 24%.

healthy d1d_{1} d2d_{2} d3d_{3} d4d_{4} d5d_{5}
healthy 64 1 13 18 3 1
d1d_{1} 1 95 0 3 1 0
d2d_{2} 9 0 83 4 2 2
d3d_{3} 24 2 8 60 5 1
d4d_{4} 1 2 1 4 90 2
d5d_{5} 4 0 4 2 0 90
Table 3: Confusion matrix considering the healthy structure, and 10% of damage damage at the first spring, at the second spring, at the third spring, at the fourth spring, and at the fifth spring.

5.3 Other datasets: further analyses

As mention before, the accuracy and efficiency of the digital twin depends on the dataset used in the training the classifier, that depends on how the input space, comprising the modal parameters and operational conditions, is explored in the off-line phase.

Table 4 shows the accuracy for different situations. The reference case has an accuracy of 93.3%, as seen in the previous section. The table shows that the accuracy tends to decrease if: the damage is lower, there are less sensors, there is more noise, and there is more uncertainty (e.g. parameter uncertainties). That is, if we have less information, or poorer information, the accuracy will decrease. And, depending on the force location and frequency, the results will change.

Cases Accuracy
Reference case 93.3%
lower damage (10%) 80.3%
less sensors (2,3,4,5,6) 74.8%
more noise (2σ\sigma) 85.3%
more uncertainty (bounds [0.90,1.10][0.90,1.10]) 81.5%
different force frequency (7000 Hz) 74.8%
different force location (first DOF) 81.0%
Table 4: Accuracy of the digital twin varying some parameters. Accuracy tends to decrease if: the damage is lower, there are less sensors, there is more noise, there is more uncertainty (e.g. parameter uncertainties). The accuracy changes if the location of the force and its frequency changes.

Figure 10 details a little more the digital twin accuracy when some parameters vary, where the red bar represents the reference case. Figure 10(a) and (b) show the accuracy increase when the damage increases, and the uncertainty bounds of the parameters of the stochastic model decrease. Figure 10(c) shows how the accuracy changes depending on the excitation frequency ωf\omega_{f}. It can be noted that the accuracy is higher when the system is excited close to the natural frequencies, 39993999 Hz and 63986398 Hz. At these frequencies the amplitude is higher and the response is more sensitive to damage comparing to the points in the neighborhood. Of course, we would not expect the system to operate close to the natural frequencies due to resonance.

Figure 10(d) shows the accuracy decrease when less sensors are taken into account. Specifically, the first sensor is the more informative for the present analysis, since removing it leads to the highest drop in the accuracy. If we take away sensors 3 and 5, the accuracy is 89.5%; if we take away sensors 2, 3 and 5 the accuracy is 80.0%. We could enter deeper in this analysis trying to optimize the number of sensors and their locations, but this is not the idea here (another possible strategy to choose sensor locations is the effective independence distribution vector [44, 45]).

Refer to caption

(a) Refer to caption(b) Refer to caption(c) Refer to caption(d)

Figure 10: Accuracy of the digital twin for different varying different parameters: (a) damage intensity, (b) parametric uncertainty, (c) excitation frequency ωf\omega_{f}, (d) sensors. The red bar is the reference case.

What if we try to extrapolate and employ the digital twin to identify damage in the physical twin under a operational condition different from the one it was trained? If this is done, we would expect a loss in its accuracy, which is what happens in our analysis, as observed in Fig. 11. For example, considering our reference configuration, trained with an excitation frequency ωf=3800\omega_{f}=3800 Hz, the accuracy of the digital twin is 93.3%. If the digital twin is trained with ωf=4000\omega_{f}=4000 Hz, the accuracy is 97.0% (a little higher), and if it is trained with ωf=3600\omega_{f}=3600 Hz, the the accuracy is 83.0% (lower). Now let the reference digital twin (obtained with ωf=3800\omega_{f}=3800 Hz) be used to classify data obtained for simulations with ωf=4000\omega_{f}=4000 and 36003600 Hz. If this is done, the accuracy drops to 93.5% (ωf=4000\omega_{f}=4000 Hz) and 77.5% (ωf=3600\omega_{f}=3600 Hz). Therefore, the generalization capacity of our digital twin is limited.

Refer to caption
Figure 11: Accuracy of the digital twin varying the frequency used for training it. The accuracy decreases when the training dataset is constructed with a different frequency of excitation ωf\omega_{f}. In blue is the accuracy when the indicated frequency is used to generate the dataset, and in red the accuracy when the dataset obtained for ωf=3800\omega_{f}=3800 is used.

In this context, it should be noticed that we are allowing the fluctuation of several parameters of the system to construct the datasets, including the frequency of excitation ωf\omega_{f}. If we fix a specific frequency (ωf=3800\omega_{f}=3800 Hz) for the training, the accuracy of the digital twin remains the same (around 93.3%). However, we argue that it is interesting to allow fluctuations of the parameters because the physical twin might be, for instance, excited by a force with a frequency different from the nominal one used in the training. Hence, these fluctuations might help the digital twin aiming at achieving a little more generalization. For example, training with the fixed frequency ωf=3800\omega_{f}=3800 Hz, and testing with the fixed frequency ωf=3600\omega_{f}=3600 Hz, the digital twin presents an accuracy of 80.5%, which is lower than 83.0% obtained when fluctuations are allowed.

Finally, Fig. 12 shows how the precision of the digital twin changes for different values of training set size. Two thirds of the samples are used for training and one third for validation. We should ensure a minimum sample size to reach convergence. The mean and the coefficient of variation (standard deviation over the mean) of the accuracy are computed with 100 simulations; and the reference case considers 200 hundred points for testing and 100 for validation, for each one of the 6 damage scenarios. This means that the training dataset has a total of 1200 rows (200×6200\times 6), and the validation dataset has a total of 600 rows (100×6100\times 6); a total of 1800 points (1200+6001200+600). If 450 points are considered to calibrate the digital twin, the mean accuracy is a little lower and the coefficient of variation is a little higher than the reference values. But, if 90 points are taken into account, the mean accuracy and the coefficient of variation are reasonably distant from the reference. It should be remarked that it is not simple to train the digital twin using data from the physical twin. Imagine testing the physical asset (e.g. wind turbine, drill string) for each damage condition, several times. Therefore, a physical-based computational model is paramount.

Refer to caption
Figure 12: Accuracy mean and coefficient of variation (standard deviation over the mean) for different sample sizes (2/3 training, 1/3 validation).

6 Concluding remarks

This paper presents a digital twin conceptual framework for a dynamic structural damage problem. A Digital twin (see section 2) is a virtual representation of a specific physical asset, built to support engineering decisions, especially in real time applications. Three important ingredients of this framework are highlighted: (i) computational model, (ii) uncertainty quantification, and (iii) calibration using data from the physical twin.

To leverage the computational strategy, a physics-based model is combined with a machine learning classifier to construct a digital twin that would be connected to the physical counterpart and support decisions.

The discrete damaged structure analyzed enabled us to construct a digital twin step-by-step, and to understand its characteristics. This knowledge might be helpful to other applications and more complex systems. The accuracy of the digital twin classifier (quadratic discriminant) decreases when: there are less sensors, there is less damage, there are less training points, and there is more uncertainty and noise. We tested Several operational conditions, and their impact on the digital twin accuracy computed. The physics-based computational model is paramount to assure interpretability and to explore different damage scenarios that could not be assessed with the physical twin. Because the physics-based model can be time consuming, training a machine learning classifier allows a fast evaluation of the physical twin in a real time operation. Future investigations include model discrepancy and prognosis [14].

Acknowledgement

We would like to acknowledge that this investigation was financed in part by the Brazilian agencies: Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES) - Finance code 001 - Grant PROEX 803/2018, Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPQ) - Grants 400933/2016-0, 302489/2016-9, and Fundação Carlos Chagas Filho de Amparo à Pesquisa do Estado do Rio de Janeiro (FAPERJ) - Grant E-26/201.572/2014.

References

  • [1] K. Schwab, The Fourth Industrial Revolution, Currency, 2017.
  • [2] D. Jones, C. Snider, A. Nassehi, J. Y, B. Hicks, Characterising the digital twin: A systematic literature review, CIRP Journal of Manufacturing Science and Technology in press (2020).
  • [3] R. Ganguli, S. Adhikari, The digital twin of discrete dynamic systems: Initial approaches and future challenges, Applied Mathematical Modelling 77 (2019) 1110–1128.
  • [4] D. Wagg, K. Worden, R. Barthorpe, P. Gardner, Digital twins: State-of-the-art and future directions for modelling and simulation in engineering dynamics applications, ASME in press (2020).
  • [5] M. A. Nielsen, Neural networks and deep learning, Determination Press, 2015.
  • [6] C. Soize, R. Ghanem, Data-driven probability concentration and sampling on manifold, Journal of Computational Physic 321 (2016) 242–258.
  • [7] D. Jones, C. Snider, A. Nassehi, J. Y, B. Hicks, A high-bias, low-variance introduction to machine learning for physicists, Physics Reports 810 (2019) 1–124.
  • [8] S. Brunton, J. Kutz, Data-driven science and engineering: machine learning, dynamical systems, and control, Cambridge University Press, 2019.
  • [9] K. Willard, X. Jia, S. Xu, M. Steinbach, V. Kumar, Integrating physics-based modeling with machine learning: A survey, ArXiv 2003.04919v3 (2020) 1–11.
  • [10] H. Kopetz, Real-Time Systems: Design Principles for Distributed Embedded Applications, 2nd Edition, Springer, 2011.
  • [11] C. Bigoni, J. S. Hesthaven, Simulation-based anomaly detection and damage localization: An application to structural health monitoring, Comput. Methods Appl. Mech. Engrg. 363 (2020) 112896.
  • [12] R. Zhao, R. Yan, Z. Chen, K. Mao, P. Wang, R. X. Gao, Deep learning and its applications to machine health monitoring, Mechanical Systems and Signal Processing 115 (2019) 213–237.
  • [13] W. Booyse, D. N. W. a, S. Heyns, Deep digital twins for detection, diagnostics and prognostics, Mechanical Systems and Signal Processing 140 (2020) 106612.
  • [14] P. M.Karve, Y. Guo, B. Kapusuzoglu, S. Mahadevan, M. A.Haileb, Digital twin approach for damage-tolerant mission planning under uncertainty, Engineering Fracture Mechanics 225 (2020) 106766.
  • [15] M. K. D. Knezevic, K. Willcox, Toward predictive digital twins via component-based reduced-order models and interpretable machine learning, AIAA Scitech Forum 0418 (2020).
  • [16] D. Bertsimas, J. Dunn, Optimal classification trees, Machine Learning 106 (7) (2017) 1039–1082.
  • [17] D. Sivia, J. Skilling, Data Analysis: A Bayesian Tutorial, Oxford University Press, 2006.
  • [18] A. Gupta, S. Arora, J. R. Westcott, Industrial Automation and Robotics, Mercury Learning and Information, 2016.
  • [19] T. Hughes, The Finite Element Method, Dover Publications, 2012.
  • [20] R. G. Lyons, Understanding Digital Signal Processing, Prentice Hall, 2010.
  • [21] C. R. Farrar, K. Worden, An introduction to structural health monitoring, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 365 (1851) (1991) 303–315.
  • [22] C. Soize, Uncertainty Quantification: An Accelerated Course with Advanced Applications in Computational Engineering, Springer, 2017.
  • [23] J. Kaipio, E. Somersalo, Statistical and Computational Inverse Problems, Springer, 2004.
  • [24] M. Friswell, J. E. Mottershead, Uncertainty Quantification: An Accelerated Course with Advanced Applications in Computational Engineering, Springer Science, 2013.
  • [25] C. Roy, W. Oberkampf, A comprehensive framework for verification, validation, and uncertainty quantification in scientific computing, Computer Methods in Applied Mechanics and Engineering 200 (25-28) (2011) 2131–2144.
  • [26] L. A. Zadeh, Fuzzy sets, Information and Controls 8 (3) (1965) 338–353.
  • [27] T. Ritto, A. Fabro, Investigation of random matrix applications on structural dynamics using a tensor decomposition, Journal of the Brazilian Society of Mechanical Sciences and Engineering 41 (352) (2019).
  • [28] C. Farhat, P. Geuzaine, G. Brown, Application of a three-field nonlinear fluid–structure formulation to the prediction of the aeroelastic parameters of an f-16 fighter, Computers & Fluids 32 (1) (2003) 3–29.
  • [29] P. Benner, S. Gugercin, K. Willcox, A survey of projection-based model reduction methods for parametric dynamical systems, SIAM Review 57 (4) (2015) 483–531.
  • [30] R. Ohayon, C. Soize, Advanced Computational Vibroacoustics: Reduced-Order Models and Uncertainty Quantification, Cambridge University Press, 2014.
  • [31] T. Ritto, F. Buezas, R. Sampaio, A new measure of efficiency for model reduction: application to a vibroimpact system, Journal of Sound and Vibration 330 (9) (2011) 1977–1984.
  • [32] B. D. Silva, D. Higdon, S. Brunton, N. Kutz, Discovery of physics from data: Universal laws and discrepancies, ArXiv 1906.07906 (2020).
  • [33] J. Han, A. Jentzen, W. E, Solving high-dimensional partial differential equations using deep learning (2018).
  • [34] A. Karpatne, G. Atluri, J. Faghmous, M. Steinbach, A. Banerjee, A. Ganguly, S. Shekhar, N. Samatova, V. Kumar, Theory-guided data science: A new paradigm for scientific discovery from data, IEEE 29 (10) (2017) 2318–2331.
  • [35] M. Raissi, P. Perdikaris, G. Karniadakis, Data-driven probability concentration and sampling on manifold, Journal of Computational Physic 378 (2016) 686–707.
  • [36] X. Jia, J. Willard, A. Karpatne, J. Read, J. Zwart, M. Steinbach, V. Kumar, Physics guided rnns for modeling dynamical systems: A case study in simulating lake temperature profiles, SIAM (2019).
  • [37] A. Daw, R. Thomas, C. Carey, J. Read, A. Appling, A. Karpatne, Physics-guided architecture (pga) of neural networks for quantifying uncertainty in lake temperature modeling, ArXiv 1911.02682 (2019).
  • [38] D. Alves, G. Daniel, H. de Castro, T. Machado, K. Cavalca, O. Gecgel, J. Dias, S. Ekwaro-Osire, Uncertainty quantification in deep convolutional neural network diagnostics of journal bearings with ovalization fault, Mechanism and Machine Theory 149 (2020) 103835.
  • [39] A. Pandey, M. Biswas, Damage detection in structures using changes in flexibility, Journal of Sound and Vibration 169 (1) (1994) 3–43176.
  • [40] R. Ribeiro, T. Ritto, Damage identification in a multi-dof system under uncertainties using optimization algorithms, Journal of Applied and Computational Mechanics 4 (4) (2018) 365–374.
  • [41] D. Castello, L. Stutz, F. Rochinha, A structural defect identification approach based on a continuum damage model, Computers & structures 180 (5-6) (2002) 417–436.
  • [42] T. Hastie, R. Tibshirani, J. Friedman, The Elements of Statistical Learning: Data Mining, Inference, and Prediction, second edition Edition, Springer, 2016.
  • [43] G. James, D. Witten, T. Hastie, R. Tibshirani, Data-driven science and engineering: machine learning, dynamical systems, and control, Springer Texts in Statistics, 2017.
  • [44] D. C. Kammer, Sensor placement for on-orbit modal identification and correlation of large space structures, Journal of Guidance Control, and Dynamics 14 (2) (1991) 251–259.
  • [45] T. Ritto, Choice of measurement locations of nonlinear structures using proper orthogonal modes and effective independence distribution vector, Shock and Vibration 697497 (2014) 1–6.

Appendix A Elementary mass and stiffness matrices of the bar model

The elementary mass and stiffness matrices of the bar model are given by (LeL_{e} is the element length):

[Mpt(e)]=ρALe[1/31/61/61/3],[M_{pt}^{(e)}]=\rho ALe\left[\begin{array}[]{cccccc}1/3&1/6\\ 1/6&1/3\\ \end{array}\right]\,, (11)
[Kpt(e)]=EALe[1111].[K_{pt}^{(e)}]=\frac{EA}{L_{e}}\left[\begin{array}[]{cccccc}1&-1\\ -1&1\\ \end{array}\right]\,. (12)

Appendix B Mass and stiffness matrices of the computational model

The mass and stiffness matrices of the computational model are given by:

[Mcm]=[m000000m000000m000000m000000m000000m],[M_{cm}]=\left[\begin{array}[]{cccccc}m&0&0&0&0&0\\ 0&m&0&0&0&0\\ 0&0&m&0&0&0\\ 0&0&0&m&0&0\\ 0&0&0&0&m&0\\ 0&0&0&0&0&m\\ \end{array}\right]\,, (13)
[Kcm]=[2kk0000k2kk0000k2kk0000k2kk0000k2kk0000kk].[K_{cm}]=\left[\begin{array}[]{cccccc}2k&-k&0&0&0&0\\ -k&2k&-k&0&0&0\\ 0&-k&2k&-k&0&0\\ 0&0&-k&2k&-k&0\\ 0&0&0&-k&2k&-k\\ 0&0&0&0&-k&k\\ \end{array}\right]\,. (14)