This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep learning for full-field ultrasonic characterization

Yang Xu1 Fatemeh Pourahmadian1,2 Jian Song1 Conglin Wang3 1 Department of Civil, Environmental & Architectural Engineering, University of Colorado Boulder, USA 2 Department of Applied Mathematics, University of Colorado Boulder, USA 3 Department of Physics, University of Colorado Boulder, USA
Abstract

This study takes advantage of recent advances in machine learning to establish a physics-based data analytic platform for distributed reconstruction of mechanical properties in layered components from full waveform data. In this vein, two logics, namely the direct inversion and physics-informed neural networks (PINNs), are explored. The direct inversion entails three steps: (i) spectral denoising and differentiation of the full-field data, (ii) building appropriate neural maps to approximate the profile of unknown physical and regularization parameters on their respective domains, and (iii) simultaneous training of the neural networks by minimizing the Tikhonov-regularized PDE loss using data from (i). PINNs furnish efficient surrogate models of complex systems with predictive capabilities via multitask learning where the field variables are modeled by neural maps endowed with (scaler or distributed) auxiliary parameters such as physical unknowns and loss function weights. PINNs are then trained by minimizing a measure of data misfit subject to the underlying physical laws as constraints. In this study, to facilitate learning from ultrasonic data, the PINNs loss adopts (a) wavenumber-dependent Sobolev norms to compute the data misfit, and (b) non-adaptive weights in a specific scaling framework to naturally balance the loss objectives by leveraging the form of PDEs germane to elastic-wave propagation. Both paradigms are examined via synthetic and laboratory test data. In the latter case, the reconstructions are performed at multiple frequencies and the results are verified by a set of complementary experiments highlighting the importance of verification and validation in data-driven modeling.

keywords:
deep learning, ultrasonic testing, data-driven mechanics, full-wavefield inversion

1 Introduction

Recent advances in laser-based ultrasonic testing has led to the emergence of dense spatiotemporal datasets which along with suitable data analytic solutions may lead to better understanding of the mechanics of complex materials and components. This includes learning of distributed mechanical properties from test data which is of interest in a wide spectrum of applications from medical diagnosis to additive manufacturing [1, 2, 3, 4, 5, 6, 7]. This work makes use of recent progress in deep learning [8, 9] germane to direct and inverse problems in partial differential equations [10, 11, 12, 13] to develop a systematic full-field inversion framework to recover the profile of pertinent physical quantities in layered components from laser ultrasonic measurements. The focus is on two paradigms, namely: the direct inversion and physics-informed neural networks (PINNs) [14, 15, 16, 17]. The direct inversion approach is in fact the authors’ rendition of elastography method [18, 19, 20] through the prism of deep learning. To this end, tools of signal processing are deployed to (a) denoise the experimental data, and (b) carefully compute the required field derivatives as per the governing equations. In parallel, the unknown distribution of PDE parameters in space-frequency are identified by neural networks which are then trained by minimizing the single-objective elastography loss. The learning process is stabilized via the Tikhonov regularization [21, 22] where the regularization parameter is defined in a distributed sense as a separate neural network which is simultaneously trained with the sought-for physical quantities. This unique exercise of learning the regularization field without a-priori estimates, thanks to neural networks, proved to be convenient, effective, and remarkably insightful in inversion of multi-fidelity experimental data.

PINNs have recently come under the spotlight for offering efficient, yet predictive, models of complex PDE systems [10] that has so far been backed by rigorous theoretical justification within the context of linear elliptic and parabolic PDEs [23]. Given the multitask nature of training for these networks and the existing challenges with modeling stiff and highly oscillatory PDEs [12, 24], much of the most recent efforts has been focused on (a) adaptive gauging of the loss function [12, 25, 26, 27, 28, 29, 13], and (b) addressing the gradient pathologies [24, 13] e.g., via learning rate annealing [30] and customizing the network architecture [11, 31, 32]. In this study, our initially austere implementations of PINNs using both synthetic and experimental waveforms led almost invariably to failure which further investigation attributed to the following impediments: (a) high-norm gradient fields due to large wavenumbers, (b) high-order governing PDEs in the case of laboratory experiments, and (c) imbalanced objectives in the loss function. These problems were further magnified by our attempts for distributed reconstruction of discontinuous PDE parameters – in the case of laboratory experiments, from contaminated and non-smooth measurements. The following measures proved to be effective in addressing some of these challenges: (i) training PINNs in a specific scaling framework where the dominant wavenumber is the reference length scale, (ii) using the wavenumber-dependent Sobolev norms in quantifying the data misfit, (iii) taking advantage of the inertia term in the governing PDEs to naturally balance the objectives in the loss function, and (iv) denoising of the experimental data prior to training.

This paper is organized as follows. Section 2 formulates the direct scattering problem related to the synthetic and laboratory experiments, and provides an overview of the data inversion logic. Section 3 presents the computational implementation of direct inversion and PINNs to reconstruct the distribution of Láme parameters in homogeneous and heterogeneous models from in-plane displacement fields. Section 4 provides a detailed account of laboratory experiments, scaling, signal processing, and inversion of antiplane particle velocity fields to recover the distribution of a physical parameter affiliated with flexural waves in thin plates. The reconstruction results are then verified by a set of complementary experiments.

2 Concept

This section provides (i) a generic formalism for the direct scattering problem pertinent to the ensuing (synthetic and experimental) full-field characterizations, and (ii) data inversion logic.

2.1 Forward scattering problem

Consider ultrasonic tests where the specimen Πd\Uppi\subset\mathbb{R}^{d}, d=2,3d=2,3, is subject to (boundary or internal) excitation over the incident surface SincΠ¯S^{\text{inc}}\!\subset\overline{\Uppi} and the induced (particle displacement or velocity) field u:Π¯×[0T]NΛ\text{\bf{u}}\colon\overline{\Uppi}\times[0\,\,T]\to\mathbb{R}^{N_{\Lambda}}\! (NΛdN_{\Lambda}\leqslant d) is captured over the observation surface SobsΠS^{\text{obs}}\!\subset\Uppi in a timeframe of length TT. Here, Π\Uppi is an open set whose closure is denoted by Π¯\overline{\Uppi}, and the sensing configuration is such that Sinc¯Sobs¯=\overline{S^{\text{inc}}}\cap\overline{S^{\text{obs}}\!}=\emptyset. In this setting, the spectrum of observed waveforms u^:Sobs×ΩNΛ\hat{\text{\bf{u}}}\colon S^{\text{obs}}\times\Omega\to\mathbb{C}^{N_{\Lambda}}\! is governed by

Λ[u^;ϑ](𝝃,ω)=𝟎,u^:=[u](𝝃,ω),𝝃Sobs,ωΩ,\Lambda[\hat{\text{\bf{u}}};{\boldsymbol{\vartheta}}](\boldsymbol{\xi},\omega)~{}=~{}\boldsymbol{0},\quad\hat{\text{\bf{u}}}~{}\colon\!\!\!=~{}\mathscr{F}[\text{\bf{u}}](\boldsymbol{\xi},\omega),\quad\boldsymbol{\xi}\in S^{\text{obs}}\!,\,\omega\in\Omega, (1)

where Λ\Lambda of size NΛ×1N_{\Lambda}\times 1 designates a differential operator in frequency-space; \mathscr{F} represents the temporal Fourier transform; ϑ{\boldsymbol{\vartheta}} of dimension Nϑ×1N_{\vartheta}\times 1 is the vector of relevant geometric and elastic parameters e.g., Lamé constants and mass density; 𝝃d\boldsymbol{\xi}\in\mathbb{R}^{d} is the position vector; and ω>0\omega>0 is the frequency of wave motion within the specified bandwidth Ω\Omega.

2.2 Dimensional platform

All quantities in (1) are rendered dimensionless by identifying ρ\rho_{\circ}, σ\sigma_{\circ}, and \ell_{\circ} as the respective reference scales [33] for mass density, elastic modulus, and length whose explicit values will be later specified.

2.3 Data inversion

Given the full waveform data u^\hat{\text{\bf{u}}} on Sobs×ΩS^{\text{obs}}\times\Omega, the goal is to identify the distribution of material properties over SobsS^{\text{obs}}. For this purpose, two reconstruction paradigms based on neural networks are adopted in this study, namely: (i) direct inversion, and (ii) physics-based neural networks. Inspired by the elastography method [18, 19], quantities of interest in (i) are identified by neural maps over Sobs×ΩS^{\text{obs}}\times\Omega that minimize a regularized measure of Λ\Lambda in (1). The neural networks in (ii), however, are by design predictive maps of the waveform data (i.e., u^\hat{\text{\bf{u}}}) obtained by minimizing the data mismatch subject to (1) as a soft or hard constraint. In this setting, the unknown properties of Λ\Lambda may be recovered as distributed parameters of the (data) network during training via multitask optimization. In what follows, a detailed description of the deployed cost functions in (i) and (ii) is provided after a brief review of the affiliated networks.

2.3.1 Waveform and parameter networks

Laser-based ultrasonic experiments furnish a dense dataset on Sobs×ΩS^{\text{obs}}\times\Omega. Based on this, multilayer perceptrons (MLPs) owing to their dense range [34] may be appropriate for approximating complex wavefields and distributed PDE parameters. Moreover, this architecture has proven successful in numerous applications within the PINN framework [15]. In this study, MLPs serve as both data and property maps where the input consists of discretized space and frequency coordinates (𝝃i,ωj)(\boldsymbol{\xi}_{i},\omega_{j}), i=1,2,,Nξi=1,2,\ldots,N_{\xi}, j=1,2,,Nωj=1,2,\ldots,N_{\omega}, as well as distinct experimental parameters, e.g., the source location, distilled as one vector τk\tau_{k} on domain 𝒯\mathscr{T} with k=1,2,,Nτk=1,2,\ldots,N_{\tau}, while the output represents waveform data 𝓓ijk=[u^,u^](𝝃i,ωj;τk)NΛ×NΛ\boldsymbol{\mathscr{D}}_{ijk}=[\mathfrak{R}{\hat{\text{\bf{u}}}},\mathfrak{I}{\hat{\text{\bf{u}}}}](\boldsymbol{\xi}_{i},\omega_{j};\tau_{k})\in\mathbb{R}^{N_{\Lambda}\!}\times\mathbb{R}^{N_{\Lambda}\!}, and/or the sought-for mechanical properties 𝓟ijn=[ϑn,ϑn](𝝃i,ωj)×\boldsymbol{\mathscr{P}}_{ijn}=[\mathfrak{R}{\vartheta}_{n},\mathfrak{I}{\vartheta}_{n}](\boldsymbol{\xi}_{i},\omega_{j})\in\mathbb{R}\times\mathbb{R}, n=1,2,,Nϑn=1,2,\ldots,N_{\vartheta}. Note that following [35], the real \mathfrak{R} and imaginary \mathfrak{I} parts of (1) and every complex-valued variable are separated such that both direct and inverse problems are reformulated in terms of real-valued quantities. In this setting, each fully-connected MLP layer with NlN_{l} neurons is associated with the forward map Υl:Nl1Nl\Upsilon_{l}\colon\mathbb{R}^{N_{l-1}}\to\mathbb{R}^{N_{l}},

Υl(𝒙l1)=tanh(𝑾l𝒙l1+𝒃l),𝒙l1Nl1,\Upsilon_{l}(\boldsymbol{x}^{l-1})~{}=~{}\text{tanh}(\boldsymbol{W}^{l}\boldsymbol{x}^{l-1}+\,\boldsymbol{b}^{l}),\quad\boldsymbol{x}^{l-1}\in\,\mathbb{R}^{N_{l-1}}, (2)

where 𝑾lNl×Nl1\boldsymbol{W}^{l}\in\mathbb{R}^{N_{l}\times N_{l-1}} and 𝒃lNl\boldsymbol{b}^{l}\in\mathbb{R}^{N_{l}} respectively denote the lthl^{\text{th}} layer’s weight and bias. Consecutive composition of Υl\Upsilon_{l} for l=1,2,,Nml=1,2,\ldots,N_{m} builds the network map wherein NmN_{m} designates the number of layers.

2.3.2 Direct inversion

Refer to caption
Figure 1: Direct inversion: (a) FFT-based spatial differentiation of the full-field data as per operator Λ\Lambda, (b) MLP-based approximation of the unknown PDE and regularization parameters (ϑ,α)({\boldsymbol{\vartheta}},\alpha) on their respective domains, and (c) training the MLPs via minimizing the elastography loss ε\mathscr{L}_{\varepsilon} according to (3).

Logically driven by the elastography method, the direct inversion approach depicted in Fig. 1 takes advantage of the leading-order physical principles underpinning the test data to recover the distribution of relevant physical quantities in space-frequency i.e., over the measurement domain. The ML-based direct inversion entails three steps: (a) spectral denoising and differentiation of (n-differentiable) waveforms u^\hat{\text{\bf{u}}} over Sobs×ΩS^{\text{obs}}\times\Omega according to the (n-th order) governing PDEs in (1), (b) building appropriate MLP maps to estimate the profile of unknown physical parameters of the forward problem and regularization parameters of the inverse solution, and (c) learning the MLPs through regularized fitting of data to the germane PDEs.

Note that synthetic datasets – generated via e.g., computer modeling or the method of manufactured solutions, may directly lend themselves to the fitting process in (c) as they are typically smooth by virtue of numerical integration or analytical form of the postulated solution. Laboratory test data, however, are generally contaminated by noise and uncertainties, and thus, spectral differentiation is critical to achieve the smoothness requirements in (c). The four-tier signal processing of experimental data follows closely that of [36, Section 3.1] which for completeness is summarized here: (1) a band-pass filter consistent with the frequency spectrum of excitation is applied to the measured time signals at every receiver point, (2) the obtained temporally smooth signals are then differentiated or integrated to obtain the pertinent field variables, (3) spatial smoothing is implemented at every snapshot in time via application of median and moving average filters followed by computing the Fourier representation of the processed waveforms in space, (4) the resulting smooth fields may be differentiated (analytically in the Fourier space) as many times as needed based on the underlying physical laws in preparation for the full-field reconstruction in step (c). It should be mentioned that the experimental data may feature intrinsic discontinuities e.g., due to material heterogeneities or contact interfaces. In this case, the spatial smoothing in (3) must be implemented in a piecewise manner after the geometric reconstruction of discontinuity surfaces in SobsS^{\text{obs}} which is quite straightforward thanks to the full-field measurements, see e.g., [36, section 3.2].

Next, the unknown PDE parameters ϑ{\boldsymbol{\vartheta}} are approximated by a fully connected MLP network ϑ:=𝒩ϑ(𝝃,ω){\boldsymbol{\vartheta}}^{\star}\colon\!\!\!=\mathscr{N}_{\boldsymbol{\vartheta}}(\boldsymbol{\xi},\omega) as per Section 2.3.1. The network is trained by minimizing the loss function

ε(u^,ϑ;α)=Λ(u^;ϑ)L2(Sobs×Ω×𝒯)NΛ2+α𝟙ϑϑL2(Sobs×Ω)Nϑ2,\mathscr{L}_{\varepsilon}(\hat{\text{\bf{u}}},{\boldsymbol{\vartheta}}^{\star};\alpha)~{}=~{}\lVert\Lambda(\hat{\text{\bf{u}}};{\boldsymbol{\vartheta}}^{\star})\rVert^{2}_{L^{2}(S^{\text{obs}}\times\hskip 0.56905pt\Omega\hskip 0.56905pt\times\mathscr{T})^{N_{\Lambda}\!}}\,+\,\lVert\alpha\boldsymbol{\mathbbm{1}}_{\vartheta}\odot{\boldsymbol{\vartheta}}^{\star}\rVert^{2}_{L^{2}(S^{\text{obs}}\times\hskip 0.56905pt\Omega)^{N_{\vartheta}}}, (3)

where 𝟙ϑ\boldsymbol{\mathbbm{1}}_{\vartheta} indicates an all-ones vector of dimension Nϑ×1N_{{\vartheta}}\times 1, and \odot designates the (element-wise) Hadamard product. Here, the PDE residual based on (1) is penalized by the norm of unknown parameters. Observe that the latter is a function of the weights and biases of the neural network which may help stabilize the MLP estimates during optimization. Such Tikhonov-type functionals are quite common in waveform tomography applications [37, 38, 39] owing to their well-established regularizing properties [21, 22]. Within this framework, α>0\mathbb{R}\ni\alpha>0 is the regularization parameter which may be determined by three means, namely: (i) the Morozov discrepancy principle [40, 41], (ii) its formulation as a (constant or distributed) parameter of the ϑ{\boldsymbol{\vartheta}}^{\star}\! network which could then be learned during training, and (iii) its independent reconstruction as a separate MLP network α:=𝒩α(𝝃,ω){{\alpha}}^{\star}\colon\!\!\!=\mathscr{N}_{\alpha}(\boldsymbol{\xi},\omega) illustrated in Fig. 1 (b) that is simultaneously trained along with ϑ{\boldsymbol{\vartheta}}^{\star} by minimizing (3). In this study, direct inversion is applied to synthetic and laboratory test data with both α=0\alpha=0 and α>0\alpha>0, based on (ii) and (iii). It was consistently observed that the regularization parameter α\alpha plays a key role in controlling the MLP estimates. This is particularly the case in situations where the field u^\hat{\text{\bf{u}}} is strongly polarized or near-zero in certain neighborhoods which brings about instability i.e., very large estimates for ϑ{\boldsymbol{\vartheta}}^{\star}\! in these areas. In light of this, all direct inversion results in this paper correspond to the case of α>0\alpha>0 identified by the MLP network α{{\alpha}}^{\star}.

2.3.3 Physics-informed neural networks

By deploying the knowledge of underlying physics, PINNs [14, 15] furnish efficient neural models of complex PDE systems with predictive capabilities. In this vein, a multitask learning process is devised according to Fig. 2 where (a) the field variable u^\hat{\text{\bf{u}}} – i.e., measured data on Sobs×Ω×𝒯S^{\text{obs}}\times\hskip 0.56905pt\Omega\hskip 0.56905pt\times\mathscr{T}, is modeled by the MLP map u^:=𝒩u^(𝝃,ω;𝝉)\hat{\text{\bf{u}}}^{\star}\colon\!\!\!=\mathscr{N}_{\hat{\text{\bf{u}}}}(\boldsymbol{\xi},\omega;\boldsymbol{\tau}) endowed with the auxiliary parameter γ(𝝃,ω;𝝉){\gamma}(\boldsymbol{\xi},\omega;\boldsymbol{\tau}) related to the loss function (4), (b) the physical unknowns ϑ{\boldsymbol{\vartheta}} could be defined either as parameters of u^\hat{\text{\bf{u}}}^{\star}\! as in Fig. 2 (i), or as a separate MLP ϑ:=𝒩ϑ(𝝃,ω){\boldsymbol{\vartheta}}^{\star}\colon\!\!\!=\mathscr{N}_{\boldsymbol{\vartheta}}(\boldsymbol{\xi},\omega) as shown in Fig. 2 (ii), and (c) learning the MLPs and affiliated parameters through minimizing a measure of data misfit subject to the governing PDEs as soft/hard constraints wherein the spatial derivatives of u^\hat{\text{\bf{u}}}^{\star} are computed via automatic differentiation [42]. It should be mentioned that in this study all MLP networks are defined on (a subset of) Sobs×Ω×𝒯S^{\text{obs}}\times\hskip 0.56905pt\Omega\hskip 0.56905pt\times\mathscr{T} where SobsΠ=S^{\text{obs}}\cap\partial\Uppi=\emptyset. Hence, the initial and boundary conditions – which could be specified as additional constraints in the loss function [15], are ignored. In this setting, the PINNs loss takes the form

ϖ(u^,ϑ|γ)=u^u^𝔑(Sobs×Ω×𝒯)NΛ2+γ𝟙ΛΛ(u^;ϑ)L2(Sobs×Ω×𝒯)NΛ2,𝔑=L2,H^ι,ιn,\mathscr{L}_{\varpi}(\hat{\text{\bf{u}}}^{\star},{\boldsymbol{\vartheta}}^{\star}|\hskip 0.56905pt{\gamma})~{}=~{}\lVert\hat{\text{\bf{u}}}\hskip 1.13809pt-\hskip 1.13809pt\hat{\text{\bf{u}}}^{\star}\rVert^{2}_{\mathfrak{N}(S^{\text{obs}}\times\hskip 0.56905pt\Omega\hskip 0.56905pt\times\mathscr{T})^{N_{\Lambda}\!}}\,\hskip 1.13809pt+\,\lVert\gamma\boldsymbol{\mathbbm{1}}_{\Lambda}\odot\Lambda(\hat{\text{\bf{u}}}^{\star};{\boldsymbol{\vartheta}}^{\star})\rVert^{2}_{L^{2}(S^{\text{obs}}\times\hskip 0.56905pt\Omega\hskip 0.56905pt\times\mathscr{T})^{N_{\Lambda}\!}},\,\,\,\mathfrak{N}\,=\,L^{2},\widehat{H}^{\iota},\,\,\,\iota\leqslant\text{n}, (4)

where 𝟙Λ\boldsymbol{\mathbbm{1}}_{\Lambda} is a NΛ×1N_{{\Lambda}}\!\times 1 vector of ones; n is the order of Λ\Lambda, and H^ι\widehat{H}^{\iota} denotes the adaptive HιH^{\iota} norm defined by

Refer to caption
Figure 2: Two logics for the physics-informed neural networks (PINNs) with distributed parameters: (i) the test data u^(𝝃,ω;𝝉)\hat{\text{\bf{u}}}(\boldsymbol{\xi},\omega;\boldsymbol{\tau}) are modeled by a MLP map, while the unknown physical parameters ϑ{\boldsymbol{\vartheta}} – on Sobs×ΩS^{\text{obs}}\times\Omega, and the loss function weight γ{{\gamma}} – on Sobs×Ω×𝒯S^{\text{obs}}\times\hskip 0.56905pt\Omega\hskip 0.56905pt\times\mathscr{T}, are defined as network parameters, and (ii) u^(𝝃,ω;𝝉)\hat{\text{\bf{u}}}(\boldsymbol{\xi},\omega;\boldsymbol{\tau}) and ϑ(𝝃,ω){\boldsymbol{\vartheta}}(\boldsymbol{\xi},\omega) are identified by separate MLPs, while γ{{\gamma}} is a parameter of 𝒩u^\mathscr{N}_{\hat{\text{\bf{u}}}}. The MLP(s) in (i) and (ii) are then trained by minimizing ϖ\mathscr{L}_{\varpi} of (4) in the space of data and PDE parameters.
H^ι:=1|𝒆|ιγ𝒆𝒆()L22+L22,𝒆=|𝒆|ξ1e1ξ2e2ξded,|𝒆|:=i=1dei.\|\cdot\|_{\widehat{H}^{\iota}}:=\sqrt{\sum_{1\hskip 0.56905pt\leqslant\hskip 0.56905pt|\boldsymbol{e}|\hskip 0.56905pt\leqslant\hskip 1.13809pt\iota}\!\!\!\!\gamma^{\boldsymbol{e}}\hskip 1.13809pt\lVert\nabla^{\boldsymbol{e}}(\hskip 0.56905pt\cdot\hskip 0.56905pt)\rVert_{L^{2}}^{2}+\lVert\hskip 1.13809pt\cdot\hskip 1.13809pt\rVert^{2}_{L^{2}}},\quad\nabla^{\boldsymbol{e}}=\frac{\partial^{|\boldsymbol{e}|}}{\partial\xi_{1}^{e_{1}}\partial\xi_{2}^{e_{2}}\!\cdot\!\hskip 1.13809pt\!\cdot\!\hskip 1.13809pt\!\cdot\!\hskip 0.56905pt\hskip 1.13809pt\partial\xi_{d}^{e_{d}}},\quad|\boldsymbol{e}|\,\colon\!\!\!=\sum_{\text{i}=1}^{d}e_{\text{i}}. (5)

Here, 𝒆:={e1,e2,ed}\boldsymbol{e}\colon\!\!\!=\{e_{1},e_{2},\ldots e_{d}\} is a vector of integers ei0e_{\text{i}}\geqslant 0. Provided that 𝒆,γ𝒆=1\forall\boldsymbol{e},\,\gamma^{\boldsymbol{e}}=1, then H^ι{\widehat{H}^{\iota}} is by definition equal to Hι{{H}^{\iota}} [43]. Note however that at high wavenumbers, Hι{{H}^{\iota}} is dominated by the highest derivatives 𝒆u^\nabla^{\boldsymbol{e}}\hat{\text{\bf{u}}}^{\star}, |𝒆|=ι|\boldsymbol{e}|=\iota, which may complicate (or even lead to the failure of) the training process due to uncontrolled error amplification by automatic differentiation particularly in earlier epochs. This issue may be addressed through proper weighting of derivatives in (5). In light of the frequency-dependent Sobolev norms in [44, 37], one potential strategy is to adopt the wavenumber-dependent weights as the following

γ𝒆=(1κ1e1κ2e2κded)2,1|𝒆|ι,\gamma^{\boldsymbol{e}}=\left(\frac{1}{\kappa_{1}^{e_{1}}\kappa_{2}^{e_{2}}\!\cdot\!\hskip 1.13809pt\!\cdot\!\hskip 1.13809pt\!\cdot\!\hskip 0.56905pt\hskip 1.13809pt\kappa_{d}^{e_{d}}}\right)^{2},\quad 1\hskip 0.56905pt\leqslant\hskip 0.56905pt|\boldsymbol{e}|\hskip 0.56905pt\leqslant\hskip 1.13809pt\iota,

wherein κi\kappa_{\text{i}} is a measure of wavenumber along ξi\xi_{\text{i}} for i=1,,d{\text{i}}=1,\ldots,d. In this setting, the weighted norms of derivatives in (5) remain approximately within the same order as the L2L^{2} norm of data misfit. Another way to automatically achieve the latter is to set the reference scale \ell_{\circ} such that κi1\kappa_{\text{i}}\!\sim\!1. Note that the H^ι{\widehat{H}^{\iota}} norms directly inform the PINNs about the “expected” field derivatives – while preventing their uncontrolled magnification. This may help stabilize the learning process as such derivatives are intrinsically involved in the PINNs loss via Λ(u^;ϑ)\Lambda(\hat{\text{\bf{u}}}^{\star};{\boldsymbol{\vartheta}}^{\star}). It should be mentioned that when 𝔑=H^ι\mathfrak{N}=\widehat{H}^{\iota} in (4), the “true” estimates for derivatives 𝒆u^\nabla^{\boldsymbol{e}}\hat{\text{\bf{u}}} may be obtained via spectral differentiation as per Section 2.3.2.

The Lagrange multiplier [45, 46] γ(𝝃,ω;𝝉)\gamma(\boldsymbol{\xi},\omega;\boldsymbol{\tau}) in (4) is critical for balancing the loss components during training. Its optimal value, however, highly depends on (a) the nature of Λ\Lambda [12], and (b) the distribution of unknown parameters ϑ\boldsymbol{\vartheta}. It should be mentioned that setting γ=1\gamma=1 led to failure in almost all of the synthetic and experimental implementations of PINNs in this study. Gauging of loss function weights has been the subject of extensive recent studies [12, 25, 47, 26, 27, 28]. One systematic approach is the adaptive SA-PINNs [12] where the multiplier γ(𝝃,ω;𝝉)\gamma(\boldsymbol{\xi},\omega;\boldsymbol{\tau}) is a distributed parameter of u^\hat{\text{\bf{u}}}^{\star} whose value is updated in each epoch according to a minimax weighting paradigm. Within this framework, the data (and parameter) networks are trained by minimizing ϖ\mathscr{L}_{\varpi} with respect to u^\hat{\text{\bf{u}}}^{\star} and ϑ{\boldsymbol{\vartheta}}^{\star}, while maximizing the loss with respect to γ\gamma as shown in Fig. 2.

Depending on the primary objective for PINNs, one may choose nonadaptive or adaptive weighting. More specifically, if the purpose is high-fidelity forward modeling via neural networks where ϑ\boldsymbol{\vartheta} is known a-priori and PINNs are intended to serve as predictive surrogate models of Λ\Lambda, then ideas rooted in constrained optimization e.g., minimax weighting is theoretically sound. However, if the inverse solution i.e., identification of ϑ(𝝃,ω)\boldsymbol{\vartheta}(\boldsymbol{\xi},\omega) from “real-world” or laboratory test data is the main goal particularly in a situation where any assumption on the smoothness of ϑ\boldsymbol{\vartheta} and/or applicability of Λ\Lambda may be (at least locally) violated e.g., due to unknown material heterogeneities or interfacial discontinuities, then trying to enforce Λ\Lambda everywhere on Sobs×Ω×𝒯S^{\text{obs}}\times\hskip 0.56905pt\Omega\hskip 0.56905pt\times\mathscr{T} (via point-wise adaptive weighting) may lead to instability and failure of data inversion. In such cases, nonadaptive weighting may be more appropriate. In light of this, in what follows, γ\gamma is a non-adaptive weight specified by taking advantage of the PDE structure to naturally balance the loss objectives.

3 Synthetic implementation

Refer to caption
Figure 3: synthetic experiments simulating plane-strain wave motion in homogeneous (top-left) and heterogeneous (bottom-left) specimens: (a) testing configuration where the model is harmonically excited at frequency ω\omega by a point source on SincS^{\text{inc}}, and the induced displacement field 𝒖\boldsymbol{u} is computed over SobsS^{\text{obs}} along ξ1\xi_{1} and ξ2\xi_{2} as shown in (b) and (c), respectively.

Full-field characterization via the direct inversion and physics-informed neural networks are examined through a set of numerical experiments. The waveform data in this section are generated via a FreeFem++ [48] code developed as part of [49].

3.1 Problem statement

Plane-strain wave motion in two linear, elastic, piecewise homogeneous, and isotropic samples is modeled according to Fig. 3 (a). On denoting the frequency of excitation by ω\omega, let r=2πωμr/ρr\ell_{r}=\frac{2\pi}{\omega}\sqrt{\mu_{r}/\rho_{r}}, ρr=1\rho_{r}=1, and μr=1\mu_{r}=1 be the reference scales for length, mass density, and stress, respectively. In this framework, both specimens are of size 1616 ×\!\times\! 1616 and uniform density ρ=1\rho=1. The first sample Π12\Uppi_{1}\subset\mathbb{R}^{2} is characterized by the constant Lamé parameters μ=1\mu_{\circ}=1 and λ=0.47\lambda_{\circ}=0.47, while the second sample Π22\Uppi_{2}\subset\mathbb{R}^{2} is comprised of four perfectly bonded homogenous components Π2j\Uppi_{2_{j}}\! of μj=j\mu_{j}=j and λj=2j/3\lambda_{j}=2j/3, j={1,2,3,4}j=\{1,2,3,4\} such that Π2¯=j=14Π2j¯\overline{\Uppi_{2}}=\bigcup_{j=1}^{4}\overline{\Uppi_{2_{j}}}. Accordingly, the shear and compressional wave speeds read cs=1\textrm{c}_{\textrm{s}}^{\circ}=1, cp=1.57\textrm{c}_{\textrm{p}}^{\circ}=1.57 in Π1\Uppi_{1}, and csj=j\textrm{c}_{\textrm{s}}^{j}=\sqrt{j}, cpj=1.63j\textrm{c}_{\textrm{p}}^{j}=1.63\sqrt{j} in Π2j\Uppi_{2_{j}}. Every numerical experiment entails an in-plane harmonic excitation at ω=3.91\omega=3.91 via a point source on SincS^{\text{inc}} (the perimeter of a 1414 ×\!\times\! 1414 square centered at the origin). The resulting displacement field 𝒖α=(u1α,u2α)\boldsymbol{u}^{\upalpha}=(u^{\upalpha}_{1},u^{\upalpha}_{2}), α=1,2\upalpha=1,2, is then computed in Πα\Uppi_{\upalpha} over SobsS^{\text{obs}} (a concentric square of dimension 88 ×\!\times\! 88) such that

μαΔ𝒖α(𝝃)+(λα+μα)𝒖α(𝝃)+ρω2𝒖α(𝝃)=δ(𝝃𝒙)𝒅,\displaystyle\mu_{\upalpha}\Delta\boldsymbol{u}^{\upalpha}(\boldsymbol{\xi})\,+\,(\lambda_{\upalpha}+\mu_{\upalpha})\nabla\nabla\cdot\boldsymbol{u}^{\upalpha}(\boldsymbol{\xi})\,+\,\rho\hskip 0.56905pt\omega^{2}\boldsymbol{u}^{\upalpha}(\boldsymbol{\xi})~{}=~{}\delta(\boldsymbol{\xi}-\boldsymbol{x})\boldsymbol{d},\quad 𝝃Πα,𝒙Sinc,\displaystyle\boldsymbol{\xi}\in{\Uppi}_{\upalpha},\boldsymbol{x}\in S^{\text{inc}}, (6)
[λα𝒖α(𝝃)𝑰2+ 2μαsym𝒖α(𝝃)]𝒏(𝝃)=𝟎,\displaystyle\big{[}\lambda_{\upalpha}\nabla\cdot\boldsymbol{u}^{\upalpha}(\boldsymbol{\xi})\boldsymbol{I}_{2}\,+\,2\mu_{\upalpha}\nabla_{\text{\tiny{sym}}}\boldsymbol{u}^{\upalpha}(\boldsymbol{\xi})\big{]}\cdot\boldsymbol{n}(\boldsymbol{\xi})~{}=~{}\boldsymbol{0},\quad 𝝃Πα,\displaystyle\boldsymbol{\xi}\in\partial{\Uppi}_{\upalpha},

where 𝒙\boldsymbol{x} and 𝒅\boldsymbol{d} respectively indicate the source location and polarization vector; 𝒏\boldsymbol{n} is the unit outward normal to the specimen’s exterior, and

{μα=μ,λα=λ,α=1μα=μj,λα=λj,α=2𝝃Π2j{1,2,3,4}.\!\left\{\begin{array}[]{l}\begin{aligned} &\!\!\mu_{\alpha}\,=\,\mu_{\circ},\,\,\lambda_{\alpha}\,=\,\lambda_{\circ},\quad&\alpha=1\!\!\!\\[0.28453pt] &\!\!\mu_{\alpha}\,=\,\mu_{j},\,\,\lambda_{\alpha}\,=\,\lambda_{j},\quad&\alpha=2\,\wedge\,\boldsymbol{\xi}\in\Uppi_{2_{j\in\{1,2,3,4\}}}\!\!\!\end{aligned}\end{array}\right..

When α=2\upalpha=2, the first of (6) should be understood as a shorthand for the set of four governing equations over Π2j\Uppi_{2_{j}}, j={1,2,3,4}j=\{1,2,3,4\}, supplemented by the continuity conditions for displacement and traction across Π2jΠ2\partial\Uppi_{2_{j}}\!\!\setminus\!\partial\Uppi_{2} as applicable.

In this setting, the generic form (1) may be identified as the following

Λ=Λα:=μαΔ+(λα+μα)+ρω2𝑰2,\displaystyle\!\!\Lambda~{}=~{}\Lambda_{\upalpha}~{}\colon\!\!\!=~{}\mu_{\upalpha}\Delta\,+\,(\lambda_{\upalpha}+\mu_{\upalpha})\nabla\nabla\cdot\,+\,\,\rho\hskip 0.56905pt\omega^{2}\boldsymbol{I}_{2}, α=1,2,\displaystyle\quad\upalpha~{}=~{}1,2, (7)
u^=𝒖α(𝝃,ω;𝝉),ϑ=[μα,λα](𝝃,ω),\displaystyle\!\!\hat{\text{\bf{u}}}~{}=~{}\boldsymbol{u}^{\upalpha}(\boldsymbol{\xi},\omega;{\boldsymbol{\tau}}),\quad\boldsymbol{\vartheta}~{}=~{}[\hskip 0.56905pt\mu_{\upalpha},\lambda_{\upalpha}](\boldsymbol{\xi},\omega), 𝝃Sobs,ωΩ,𝝉𝒯,\displaystyle\quad\boldsymbol{\xi}\in S^{\text{obs}}\!,\,\omega\in\Omega,{\boldsymbol{\tau}}\in\mathscr{T},

wherein 𝑰2\boldsymbol{I}_{2}\! is the second-order identity tensor; 𝝉=(𝒙,𝒅)Sinc×1=𝒯{\boldsymbol{\tau}}=(\boldsymbol{x},\boldsymbol{d})\in S^{\text{inc}}\times\mathscr{B}_{1}=\mathscr{T} with 1\mathscr{B}_{1} denoting the unit circle of polarization directions. Note that ρ\rho is treated here as a known parameter.

In the numerical experiments, SincS^{\text{inc}} (respSobsS^{\text{obs}}) is discretized by a uniform grid of 32 (resp. 50 ×\!\times\! 50) points, while Ω\Omega and 1\mathscr{B}_{1} are respectively sampled at ω=3.91\omega=3.91 and 𝒅=(1,0)\boldsymbol{d}=(1,0).

All inversions in this study are implemented within the PyTorch framework [50].

3.2 Direct inversion

The three-tier logic of Section 2.3.2 is employed to reconstruct the distribution of μα\mu_{\upalpha} and λα\lambda_{\upalpha}, α=1,2\upalpha=1,2, over SobsS^{\text{obs}}, entailing: (a) spectral differentiation of the displacement field 𝒖α\boldsymbol{u}^{\upalpha} in order to compute Δ𝒖α\Delta\boldsymbol{u}^{\upalpha} and 𝒖α\nabla\nabla\cdot\boldsymbol{u}^{\upalpha} as per (6), (b) construction of three positive-definite MLP networks μ\mu^{\star}, λ\lambda^{\star}, and α\alpha^{\star}; each of which is comprised of one hidden layer of 64 neurons, and (c) training the MLPs by minimizing ε\mathscr{L}_{\varepsilon} as in (3) and (7) by way of the ADAM algorithm [51]. To avoid near-boundary errors affiliated with the one-sided FFT differentiation in Δ𝒖α\Delta\boldsymbol{u}^{\upalpha} and 𝒖α\nabla\nabla\cdot\boldsymbol{u}^{\upalpha}, a concentric 40×4040\times 40 subset of collocation points sampling SobsS^{\text{obs}} is deployed for training purposes. It should also be mentioned that in the heterogeneous case, i.e., α=2\upalpha=2, the discontinuity of derivatives across Π2j{1,2,3,4}\partial\Uppi_{2_{j\in\{1,2,3,4\}}}\! calls for piecewise spectral differentiation. According to Section 2.3.1, the input to 𝒫=𝒩𝒫(𝝃,ω)\mathscr{P}^{\star}=\mathscr{N}_{\mathscr{P}}(\boldsymbol{\xi},\omega), 𝒫=μ,λ\mathscr{P}=\mu,\lambda, and α=𝒩α(𝝃,ω)\alpha^{\star}=\mathscr{N}_{\alpha}(\boldsymbol{\xi},\omega) is of size NξNτ×Nω=1600Ns×1N_{\xi}N_{\tau}\times N_{\omega}=1600N_{s}\times 1 where Ns32N_{s}\leqslant 32 is the number of simulations i.e., source locations used to generate distinct waveforms for training. In this setting, since the physical quantities of interest are independent of 𝝉{\boldsymbol{\tau}}, the real-valued output of MLPs is of dimension 1600×11600\times 1 furnishing a local estimate of the Láme and regularization parameters at the specified sampling points on SobsS^{\text{obs}}. Each epoch makes use of the full dataset and the learning rate is 0.0050.005.

In this work, the reconstruction error is measured in terms of the normal misfit

Ξ(q)=qqL2qL,\Xi(\text{q}^{\star})~{}=~{}\frac{\parallel\!\text{q}^{\star}\hskip 1.13809pt-\,\text{q}\hskip 0.56905pt\!\parallel_{L^{2}}}{\parallel\!\text{q}\hskip 0.56905pt\!\parallel_{L^{\infty}}}, (8)

where q\text{q}^{\star} is an MLP estimate for a quantity with the “true” value q.

Let SincS^{\text{inc}} be sampled at one point i.e., Ns=1N_{s}=1 so that a single forward simulation in Πα\Uppi_{\upalpha}, α=1,2\upalpha=1,2, generates the training dataset. The resulting reconstructions are shown in Figs. 4 and 5. It is evident from both figures that the single-source reconstruction fails at the loci of near-zero displacement which may explain the relatively high values of the recovered regularization parameter α\alpha^{\star}. Table 1 details the true values as well as mean and standard deviation of the reconstructed Láme distributions ϑ=(μ,λ)\boldsymbol{\vartheta}^{\star}=(\mu^{\star},\lambda^{\star}) in Π1\Uppi_{1} (respΠ2j\Uppi_{2_{j}} for j=1,2,3,4j=1,2,3,4) according to Fig. 4 (resp. Fig. 5).

This problem may be addressed by enriching the training dataset e.g., via increasing NsN_{s}. Figs. 6 and 7 illustrate the reconstruction results when SincS^{\text{inc}} is sampled at Ns=5N_{s}=5 source points. The mean and standard deviation of the reconstructed distributions are provided in Table 2. It is worth noting that in this case the identified regularization parameter α\alpha^{\star} assumes much smaller values – compared to that of Figs. 4 and 5. This is closer to the scale of computational errors in the forward simulations.

To examine the impact of noise on the reconstruction, the multisource dataset used to generate Figs. 6 and 7 are perturbed with 5%5\% white noise. The subsequent direct inversions from noisy data are displayed in Figs. 8 and 9, and the associated statistics are presented in Table 3. Note that spectral differentiation as the first step in direct inversion plays a critical role in denoising the waveforms, and subsequently regularizing the reconstruction process. This may substantiate the low magnitude of MLP-recovered α\alpha^{\star} in the case of noisy data in Figs. 8 and 9. The presence of noise, nonetheless, affects the magnitude and thus composition of terms in the Fourier representation of the processed displacement fields in space which is used for differentiation. This may in turn lead to the emergence of fluctuations in the reconstructed fields.

Refer to caption
Figure 4: Direct inversion of the Láme parameters in Π1\Uppi_{1} using noiseless data from a single source: (a) MLP-predicted distributions μ\mu^{\star} and λ\lambda^{\star}, (b) reconstruction error (8) with respect to the true values μ=1\mu_{\circ}=1 and λ=0.47\lambda_{\circ}=0.47, (c) MLP-recovered distribution of the regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e} in the log=log10\log=\log_{10} scale.
Refer to caption
Figure 5: Direct inversion of the Láme parameters in Π2\Uppi_{2} using noiseless data from a single source: (a) MLP-predicted distributions μ\mu^{\star} and λ\lambda^{\star}, (b) reconstruction error (8) with respect to the true values μj=j\mu_{j}=j and λj=2j/3\lambda_{j}=2j/3, j={1,2,3,4}j=\{1,2,3,4\}, (c) MLP-recovered regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e}.
Table 1: Mean 𝒟\langle\hskip 0.56905pt\cdot\hskip 0.56905pt\rangle_{\mathscr{D}} and standard deviation σ(|𝒟)\upsigma(\hskip 0.56905pt\cdot\hskip 0.56905pt|_{\mathscr{D}}) of the reconstructed Láme distributions in 𝒟=Π1,Π2j=1,2,3,4\mathscr{D}=\Uppi_{1},\Uppi_{2_{j=1,2,3,4}}. Here, the direct inversion is applied to noiseless data from a single source as shown in Figs. 4 and 5.
𝒟\mathscr{D} Π1\Uppi_{1} Π21\Uppi_{2_{1}} Π22\Uppi_{2_{2}} Π23\Uppi_{2_{3}} Π24\Uppi_{2_{4}}
μ\mu μ=1\mu_{\circ}=1 μ1=1\mu_{1}=1 μ2=2\mu_{2}=2 μ3=3\mu_{3}=3 μ4=4\mu_{4}=4
μ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\mu^{\star}\rangle_{\mathscr{D}} ​​0.9980.998​​ ​​0.9910.991​​ ​​1.9831.983​​ ​​2.8252.825​​ ​​3.8353.835​​
σ(μ|𝒟)\upsigma(\mu^{\star}|_{\mathscr{D}}) ​​0.0240.024​​ ​​0.0830.083​​ ​​0.1820.182​​ ​​0.4410.441​​ ​​0.3250.325​​
λ\lambda λ=0.47\lambda_{\circ}=0.47 λ1=0.67\lambda_{1}=0.67 λ2=1.33\lambda_{2}=1.33 λ3=2\lambda_{3}=2 λ4=2.66\lambda_{4}=2.66
λ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\lambda^{\star}\rangle_{\mathscr{D}} ​​0.3760.376​​ ​​0.6150.615​​ ​​0.8500.850​​ ​​1.7461.746​​ ​​1.4121.412​​
σ(λ|𝒟)\upsigma(\lambda^{\star}|_{\mathscr{D}}) ​​0.1280.128​​ ​​0.1610.161​​ ​​0.3990.399​​ ​​0.4860.486​​ ​​0.8640.864​​
Refer to caption
Figure 6: Direct inversion of the Láme parameters in Π1\Uppi_{1} using noiseless data from five distinct simulations: (a) MLP-predicted distributions μ\mu^{\star} and λ\lambda^{\star}, (b) reconstruction error (8) with respect to the true values μ=1\mu_{\circ}=1 and λ=0.47\lambda_{\circ}=0.47, (c) MLP-recovered regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e}.
Refer to caption
Figure 7: Direct inversion of the Láme parameters in Π2\Uppi_{2} using five noiseless datasets: (a) MLP-predicted distributions μ\mu^{\star} and λ\lambda^{\star}, (b) reconstruction error (8) with respect to the true values μj=j\mu_{j}=j and λj=2j/3\lambda_{j}=2j/3, j={1,2,3,4}j=\{1,2,3,4\}, (c) MLP-recovered regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e}.
Table 2: Mean and standard deviation of the reconstructed Láme distributions from five distinct noiseless datasets according to Figs. 6 and 7.
𝒟\mathscr{D} Π1\Uppi_{1} Π21\Uppi_{2_{1}} Π22\Uppi_{2_{2}} Π23\Uppi_{2_{3}} Π24\Uppi_{2_{4}}
μ\mu 11 11 22 33 44
μ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\mu^{\star}\rangle_{\mathscr{D}} 1.0001.000 0.9990.999 2.0032.003 2.9992.999 3.9993.999
σ(μ|𝒟)\upsigma(\mu^{\star}|_{\mathscr{D}}) ​​0.0010.001​​ ​​0.0120.012​​ ​​0.0110.011​​ ​​0.0120.012​​ ​​0.0160.016​​
λ\lambda 0.470.47 0.670.67 1.331.33 22 2.662.66
λ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\lambda^{\star}\rangle_{\mathscr{D}} ​​0.4640.464​​ ​​0.6600.660​​ ​​1.3021.302​​ ​​1.9971.997​​ ​​2.6352.635​​
σ(λ|𝒟)\upsigma(\lambda^{\star}|_{\mathscr{D}}) ​​0.0120.012​​ ​​0.0390.039​​ ​​0.0710.071​​ ​​0.0480.048​​ ​​0.0680.068​​
Refer to caption
Figure 8: Direct inversion of the Láme parameters in Π1\Uppi_{1} using five datasets perturbed with 5% white noise: (a) MLP-predicted distributions μ\mu^{\star} and λ\lambda^{\star}, (b) reconstruction error (8) with respect to the true values μ=1\mu_{\circ}=1 and λ=0.47\lambda_{\circ}=0.47, (c) MLP-recovered regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e}.
Refer to caption
Figure 9: Direct inversion of the Láme parameters in Π2\Uppi_{2} using five datasets perturbed with 5% white noise: (a) MLP-predicted distributions μ\mu^{\star} and λ\lambda^{\star}, (b) reconstruction error (8) with respect to the true values μj=j\mu_{j}=j and λj=2j/3\lambda_{j}=2j/3, j={1,2,3,4}j=\{1,2,3,4\}, (c) MLP-recovered regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e}.
Table 3: Mean and standard deviation of the reconstructed Láme distributions from noisy data according to Figs. 8 and 9.
𝒟\mathscr{D} Π1\Uppi_{1} Π21\Uppi_{2_{1}} Π22\Uppi_{2_{2}} Π23\Uppi_{2_{3}} Π24\Uppi_{2_{4}}
μ\mu 11 11 22 33 44
μ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\mu^{\star}\rangle_{\mathscr{D}} 1.0011.001 1.0021.002 2.0052.005 2.9962.996 3.9963.996
σ(μ|𝒟)\upsigma(\mu^{\star}|_{\mathscr{D}}) ​​0.0050.005​​ ​​0.0160.016​​ ​​0.0350.035​​ ​​0.0540.054​​ ​​0.0880.088​​
λ\lambda 0.470.47 0.670.67 1.331.33 22 2.662.66
λ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\lambda^{\star}\rangle_{\mathscr{D}} ​​0.4620.462​​ ​​0.6500.650​​ ​​1.2631.263​​ ​​2.0062.006​​ ​​2.6542.654​​
σ(λ|𝒟)\upsigma(\lambda^{\star}|_{\mathscr{D}}) ​​0.0420.042​​ ​​0.0510.051​​ ​​0.2250.225​​ ​​0.1820.182​​ ​​0.3000.300​​

3.3 Physics-informed neural networks

The learning process of Section 2.3.3 is performed as follows: (a) the MLP network 𝒖α=𝒩𝒖α(𝝃,ω,𝒙|γ,ϑ){\boldsymbol{u}^{\upalpha}}^{\star}=\mathscr{N}_{\boldsymbol{u}^{\upalpha}}(\boldsymbol{\xi},\omega,\boldsymbol{x}\hskip 1.13809pt|\hskip 1.13809pt\gamma,\boldsymbol{\vartheta}^{\star}) endowed with the positive-definite parameters γ\gamma and ϑ=(μ,λ)\boldsymbol{\vartheta}^{\star}=(\mu^{\star},\lambda^{\star}) is constructed such that the input 𝒙\boldsymbol{x} labels the source location and the auxiliary weight γ{\gamma} is a nonadaptive scaler, (b) μ\mu^{\star} and λ\lambda^{\star} may be specified as scaler or distributed parameters of the network according to Fig. 2 (i), and (c) 𝒖α{\boldsymbol{u}^{\upalpha}}^{\star} is trained by minimizing ϖ\mathscr{L}_{\varpi} in (4) via the ADAM optimizer using the synthetic waveforms of Section 3.1. Reconstructions are performed on the same set of collocation points sampling Sobs×Ω×𝒯S^{\text{obs}}\!\times\Omega\times\!\mathscr{T} as in Section 3.2. Accordingly, the input to 𝒖α{\boldsymbol{u}^{\upalpha}}^{\star} is of size Nξ×Nω×Nτ=1600×1×NsN_{\xi}\!\times\!N_{\omega}\!\times\!N_{\tau}=1600\!\times\!1\!\times\!N_{s}, while its output is of dimension (1600×1×Ns)2(1600\!\times\!1\!\times\!N_{s})^{2} modeling the displacement field along ξ1\xi_{1} and ξ2\xi_{2} in the sampling region. Similar to Section 3.2, each epoch makes use of the full dataset for training and the learning rate is 0.0050.005. The PyTorch implementation of PINNs in this section is accomplished by building upon the available codes on the Github repository [52].

The MLP network 𝒖1=𝒖1(𝝃,ω,𝒙|γ,ϑ){\boldsymbol{u}^{1}}^{\star}={\boldsymbol{u}^{1}}^{\star}(\boldsymbol{\xi},\omega,\boldsymbol{x}\hskip 1.13809pt|\hskip 1.13809pt\gamma,\boldsymbol{\vartheta}^{\star}) with three hidden layers of respectively 2020, 4040, and 2020 neurons is employed to map the displacement field 𝒖1{\boldsymbol{u}^{1}} (in Π1\Uppi_{1}) associated with a single point source of frequency ω=3.91\omega=3.91 at 𝒙=𝒙1Sinc\boldsymbol{x}=\boldsymbol{x}_{1}\in S^{\text{inc}}. The Láme constants are defined as the unknown scaler parameters of the network i.e., ϑ={μ,λ}\boldsymbol{\vartheta}^{\star}=\{\mu^{\star},\lambda^{\star}\}, and the Lagrange multiplier γ\gamma is specified per the following argument. Within the dimensional framework of this section and with reference to (7), observe that on setting γ=1ρω2\gamma=\frac{1}{\rho\omega^{2}} (i.e., γ=0.065\gamma=0.065), both (the PDE residue and data misfit) components of the loss function ϖ\mathscr{L}_{\varpi} in 4 emerge as some form of balance in terms of the displacement field. This may naturally facilitate maintaining of the same scale for the loss terms during training, and thus, simplifying the learning process by dispensing with the need to tune an additional parameter γ\gamma. Keep in mind that the input to 𝒖1{\boldsymbol{u}^{1}}^{\star} is of size 1600×1×11600\!\times\!1\!\times\!1, while its output is of dimension (1600×1×1)2(1600\!\times\!1\!\times\!1)^{2}. In this setting, the training objective is two-fold: (a) construction of a surrogate map for 𝒖1{\boldsymbol{u}^{1}}, and (b) identification of μ\mu^{\star} and λ\lambda^{\star}.

Fig. 10 showcases (i) the accuracy of PINN estimates based on noiseless data in terms of the vertical component of displacement field u21{u}^{1}_{2} in Π1\Uppi_{1}, and (ii) the performance of automatic differentiation [42] in capturing the field derivatives in terms of components that appear in the governing PDE 7 i.e., u2,ij1=2u21/(ξiξj){u}^{1}_{2,ij}={\partial^{2}{u}^{1}_{2}}/(\partial\xi_{i}\partial\xi_{j}), i,j=1,2i,j=1,2. The comparative analysis in (ii) is against the spectral derivates of FEM fields according to Section 2.3.2. It is worth noting that similar to Fourier-based differentiation, the most pronounced errors in automatic differentiation occur in the near-boundary region i.e., the support of one-sided derivatives. It is observed that the magnitude of such discrepancies may be reduced remarkably by increasing the number of epochs. Nonetheless, the loci of notable errors remain at the vicinity of specimen’s external boundary or internal discontinuities such as cracks or material interfaces. Fig. 10 is complemented with the reconstruction

Refer to caption
Figure 10: PINN vs. FEM maps of vertical displacement and its derivatives in Π1\Uppi_{1}: (a) MLP estimates, from noiseless data, for {u21,u12,11,u12,22,u12,12}\{{u_{2}^{1}}^{\star}\!,{u^{1}}^{\star}_{\!\!\!2,11},{u^{1}}^{\star}_{\!\!\!2,22},{u^{1}}^{\star}_{\!\!\!2,12}\} wherein the derivatives u12,ij{u^{1}}^{\star}_{\!\!\!2,ij}, i,j=1,2i,j=1,2, are obtained by automatic differentiation, (b) FEM displacement solution and its spectral derivatives for {u21,u2,111,u2,221,u2,121}\{{u}_{2}^{1},{u}^{1}_{2,11},{u}^{1}_{2,22},{u}^{1}_{2,12}\}, and (c) normal misfit 8 between (a) and (b).
Refer to caption
Figure 11: PINN reconstruction of Láme constants in the homogeneous plate Π1\Uppi_{1} from noiseless data: (a) μ\mu^{\star} vs. number of epochs NeN_{e}, (b) λ\lambda^{\star} vs. NeN_{e}, and (c) total loss ϖ\mathscr{L}_{\varpi} and its components (the PDE residue and data misfit) vs. NeN_{e} in log scale.
Refer to caption
Figure 12: PINN reconstruction of Láme constants in Π1\Uppi_{1} from noisy data: (a) μ\mu^{\star} vs. number of epochs NeN_{e}, (b) λ\lambda^{\star} vs. NeN_{e}, and (c) total loss ϖ\mathscr{L}_{\varpi} and its components (the PDE residue and data misfit) vs. NeN_{e} in log scale.

results of Fig. 11 indicating (μ,λ)=(1.000,0.486)(\mu^{\star},\lambda^{\star})=(1.000,0.486) for the homogenous specimen Π1\Uppi_{1} with the true Láme constants (μ,λ)=(1,0.47)(\mu_{\circ},\lambda_{\circ})=(1,0.47). The impact of noise on training is examined by perturbing the noiseless data related to Fig. 10 with 5%5\% white noise, which led to (μ,λ)=(0.999,0.510)(\mu^{\star},\lambda^{\star})=(0.999,0.510) as shown in Fig. 12.

Next, the PINN 𝒖2=𝒖2(𝝃,ω,𝒙|ϑ){\boldsymbol{u}^{2}}^{\star}={\boldsymbol{u}^{2}}^{\star}(\boldsymbol{\xi},\omega,\boldsymbol{x}\hskip 1.13809pt|\hskip 1.13809pt\boldsymbol{\vartheta}^{\star}) with three hidden layers of respectively 120120, 120120, and 8080 neurons is created to reconstruct (i) displacement field 𝒖2{\boldsymbol{u}^{2}} in the heterogeneous specimen Π2\Uppi_{2}, and (ii) distribution of the Láme parameters over the observation surface. In this vein, synthetic waveform data associated with five point sources {𝒙i}Sinc\{\boldsymbol{x}_{i}\}\in S^{\text{inc}}, i=1,2,,5i=1,2,\ldots,5 at ω=3.91\omega=3.91 is used for training. Here, ϑ\boldsymbol{\vartheta}^{\star} is the network’s unknown distributed parameter, of dimension (40×40)2(40\!\times\!40)^{2}, and the nonadaptive scaler weight γ=0.065\gamma=0.065 in light of the sample’s uniform density ρ=1\rho=1. In this setting, the input to 𝒖2{\boldsymbol{u}^{2}}^{\star} is of size 1600×1×51600\!\times\!1\!\times\!5, while its output is of dimension (1600×1×5)2(1600\!\times\!1\!\times\!5)^{2}. Fig. 13 provides a comparative analysis between the FEM and PINN maps of horizontal displacement u21{u}^{1}_{2} in Π2\Uppi_{2} and its spatial derivatives computed by spectral and automatic differentiation respectively.

Refer to caption
Figure 13: PINN vs. FEM maps of horizontal displacement and its derivatives in Π2\Uppi_{2}: (a) PINN estimates, from noiseless data, for {u12,u21,11,u21,22,u21,12}\{{u_{1}^{2}}^{\star}\!,{u^{2}}^{\star}_{\!\!\!1,11},{u^{2}}^{\star}_{\!\!\!1,22},{u^{2}}^{\star}_{\!\!\!1,12}\} wherein the derivatives u21,ij{u^{2}}^{\star}_{\!\!\!1,ij}, i,j=1,2i,j=1,2, are obtained by automatic differentiation, (b) FEM displacement solution and its spectral derivatives for {u12,u1,112,u1,222,u1,122}\{{u}_{1}^{2},{u}^{2}_{1,11},{u}^{2}_{1,22},{u}^{2}_{1,12}\}, and (c) normal misfit 8 between (a) and (b).
Refer to caption
Figure 14: PINN reconstruction of Láme parameters in Π2\Uppi_{2} using five noiseless datasets: (a) PINN-predicted distributions μ\mu^{\star} and λ\lambda^{\star}, (b) reconstruction error (8) with respect to the true values μj=j\mu_{j}=j and λj=2j/3\lambda_{j}=2j/3, j={1,2,3,4}j=\{1,2,3,4\}, (c) total loss ϖ\mathscr{L}_{\varpi} and its components (the PDE residue and data misfit) vs. NeN_{e} in log scale.
Table 4: Mean and standard deviation of the PINN-reconstructed Láme distributions from five distinct noiseless datasets according to Fig. 14.
𝒟\mathscr{D} Π21\Uppi_{2_{1}} Π22\Uppi_{2_{2}} Π23\Uppi_{2_{3}} Π24\Uppi_{2_{4}}
μ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\mu^{\star}\rangle_{\mathscr{D}} 0.9750.975 1.9731.973 2.9412.941 . 3.9183.918
σ(μ|𝒟)\upsigma(\mu^{\star}|_{\mathscr{D}}) ​​0.0540.054​​ ​​0.1230.123​​ ​​0.1350.135​​ ​​0.2260.226​​
λ𝒟\hskip 1.13809pt\langle\hskip 0.56905pt\lambda^{\star}\rangle_{\mathscr{D}} ​​0.6860.686​​ ​​1.2501.250​​ ​​2.0452.045​​ ​​2.0652.065​​
σ(λ|𝒟)\upsigma(\lambda^{\star}|_{\mathscr{D}}) ​​0.2470.247​​ ​​0.4000.400​​ ​​0.5200.520​​ ​​0.8570.857​​

The PINN-reconstructed distribution of PDE parameters is illustrated in Fig. 14 whose statistics is detailed in Table 4. It is worth mentioning that the learning process is repeated for a suit of weights γ={0.01,0.025,0.1,0.25,0.5,1.5,2,5,10,15}\gamma=\{0.01,0.025,0.1,0.25,0.5,1.5,2,5,10,15\}. In all cases, the results are either quite similar or worse than that of Figs. 13 and 14.

4 Laboratory implementation

This section examines the performance of direct inversion and PINNs for full-field ultrasonic characterization in a laboratory setting. In what follows, experimental data are processed prior to inversion as per Section 2.3.2 which summarizes the detailed procedure in [36]. To verify the inversion results, quantities of interest are also reconstructed through dispersion analysis, separately, from a set of auxiliary experiments.

4.1 Test set-up

Refer to caption
Figure 15: Test set-ups for ultrasonic full-field characterization: (a) an Al plate Π1exp\Uppi_{1}^{{}^{\text{exp}}}\! is subject to antiplane shear waves at 165165 kHz by a piezoelectric transducer; the out-of-plane particle velocity field is then captured by a laser Doppler vibrometer scanning on a robot over the observation surface, and (b) a Ti-St-Br plate Π2exp\Uppi_{2}^{{}^{\text{exp}}}\! undergoes a similar test at 8080 kHz and 300300 kHz.

Experiments are performed on two (homogeneous and heterogeneous) specimens: Π1exp\Uppi_{1}^{{}^{\text{exp}}}\! which is a 2727 cm ×27\!\times 27 cm ×1.5\!\times 1.5 mm sheet of T6 6061 aluminum, and Π2exp\Uppi_{2}^{{}^{\text{exp}}}\! composed of (a) 55 cm ×27\!\times\hskip 0.56905pt27 cm ×1.5\!\times 1.5 mm sheet of Grade 2 titanium, (b) 2.52.5 cm ×27\!\times\hskip 0.56905pt27 cm ×1.5\!\times 1.5 mm sheet of 4130 steel, and (c) 55 cm ×27\!\times\hskip 0.56905pt27 cm ×1.5\!\times 1.5 mm sheet of 260-H02 brass, connected via metal epoxy. For future reference, the density ρμ\uprho_{\mu}, Young’s modulus Eμ\textsf{E}_{\mu}, and Poisson’s ratio νμ\nu_{\mu} for μ={Al, Ti, St, Br}\mu=\{\text{Al, Ti, St, Br}\} are listed in Table 5 as per the manufacturer.

Ultrasonic experiments on both samples are performed in a similar setting in terms of the sensing configuration and illuminating wavelet. In both cases, the specimen is excited by an antiplane shear wave from a designated source location SincS^{\text{inc}}, shown in Fig. 15, by a 0.50.5 MHz p-wave piezoceramic transducer (V101RB by Olympus Inc.). The incident signal is a five-cycle burst of the form

H(𝖿𝖼𝗍)H(5𝖿𝖼𝗍)sin(0.2π𝖿𝖼𝗍)sin(2π𝖿𝖼𝗍),H({\sf f_{c}t})\,H(5\!-\!{\sf f_{c}t})\,\sin\big{(}0.2\pi{\sf f_{c}t}\big{)}\,\sin\big{(}2\pi{\sf f_{c}t}\big{)}, (9)

where HH denotes the Heaviside step function, and the center frequency 𝖿𝖼{\sf f_{c}}\! is set at 165165 kHz (resp{80,300}\{80,300\} kHz) in Π1exp\Uppi_{1}^{{}^{\text{exp}}}\! (respΠ2exp\Uppi_{2}^{{}^{\text{exp}}}). The induced wave motion is measured in terms of the particle velocity 𝗏β{\sf{v}^{\upbeta}}, β=1,2\upbeta=1,2, on the scan grids 𝒢β\mathscr{G}_{\upbeta} sampling SobsS^{\text{obs}} where SobsSinc=SobsΠβexp=S^{\text{obs}}\cap S^{\text{inc}}=S^{\text{obs}}\cap\partial\Uppi_{\upbeta}^{{}^{\text{exp}}}\!=\emptyset. A laser Doppler vibrometer (LDV) which is mounted on a 2D robotic translation frame (for scanning) is deployed for measurements. The VibroFlex Xtra VFX-I-120 LDV system by Polytec Inc. is capable of capturing particle velocity within the frequency range DC24\sim\text{DC}-24 MHz along the laser beam which in this study is normal to the specimen’s surface.

The scanning grid 𝒢1Π1exp\mathscr{G}_{1}\subset\Uppi_{1}^{{}^{\text{exp}}}\! is identified by a 22 cm ×2\!\times\hskip 0.56905pt2 cm square sampled by 100×100100\!\times\!100 uniformly spaced measurement points. This amounts to a spatial resolution of 0.20.2 mm in both spatial directions. In parallel, 𝒢2Π2exp\mathscr{G}_{2}\subset\Uppi_{2}^{{}^{\text{exp}}}\! is a 2.52.5 cm ×7.5\!\times\hskip 0.56905pt7.5 cm rectangle positioned according to Fig. 15 (b) and sampled by a uniform grid of 180×60180\!\times\!60 scan points associated with the spatial resolution of 0.420.42 mm. At every scan point, the data acquisition is conducted for a time period of 400400 μ\mus at the sampling rate of 250250 MHz. To minimize the impact of optical and mechanical noise in the system, the measurements are averaged over an ensemble of 80 realizations at each scan point. Bear in mind that both the direct inversion and PINNs deploy the spectra of normalized velocity fields vobsv^{\text{obs}} for data inversion. Such distributions of out-of-plane particle velocity at 165165 kHz (resp8080 kHz) in Π1exp\Uppi_{1}^{{}^{\text{exp}}}\! (respΠ2exp\Uppi_{2}^{{}^{\text{exp}}}) is displayed in Fig. 15.

It should be mentioned that in the above experiments, the magnitude of measured signals in terms of displacement is of O(nm)O(\text{nm}) so that it may be appropriate to assume a linear regime of propagation. The nature of antiplane wave motion is dispersive nonetheless. Therefore, to determine the relevant length scales in each component, the associated dispersion curves are obtained as in Fig. 19 via a set of complementary experiments described in Section 4.4.1. Accordingly, for excitations of center frequency {𝖿𝖼𝟣,𝖿𝖼𝟤,𝖿𝖼𝟥}={165,80,300}\{{\sf f_{c_{1}}},{\sf f_{c_{2}}},{\sf f_{c_{3}}}\}=\{165,80,300\} kHz, the affiliated phase velocity cμ\textsf{c}_{\mu} and wavelength λμ\uplambda_{\mu} for μ={Al, Ti, St, Br}\mu=\{\text{Al, Ti, St, Br}\} is identified in Table 6.

4.2 Dimensional framework

On recalling Section 2.2, let r:=λAl=0.01\ell_{r}\colon\!\!\!=\uplambda_{\text{Al}}=0.01 m, μr:=EAl=68.9\mu_{r}\colon\!\!\!=\textsf{E}_{\text{Al}}=68.9 GPA, and ρr:=ρAl=2700\rho_{r}\colon\!\!\!=\uprho_{\text{Al}}=2700 kg/m3 be the reference scales for length, stress, and mass density, respectively. In this setting, the following maps take the physical quantities to their dimensionless values

(ρμ,Eμ,νμ)(ρμ,Eμ,νμ):=(1ρrρμ,1μrEμ,νμ),\displaystyle(\uprho_{\mu},\textsf{E}_{\mu},\nu_{\mu})~{}\rightarrow~{}(\rho_{\mu},E_{\mu},\nu_{\mu}):=\big{(}\dfrac{1}{\rho_{r}}\uprho_{\mu},\dfrac{1}{\mu_{r}}\textsf{E}_{\mu},\nu_{\mu}\big{)}, μ={Al, Ti, St, Br},\displaystyle\!\!\!\!\!\!\mu=\{\text{Al, Ti, St, Br}\}, (10)
(𝖿𝖼ι,λμ,cμ)(fcι,λμ,cμ):=(rρrμr𝖿𝖼ι,1rλμ,ρrμrcμ),\displaystyle({\sf f_{c_{\iota}}},\uplambda_{\mu},\textsf{c}_{\mu})~{}\rightarrow~{}({f_{c_{\iota}}},\lambda_{\mu},{c}_{\mu}):=\big{(}\ell_{r}\sqrt{\frac{\rho_{r}}{\mu_{r}}}\hskip 1.13809pt{\sf f_{c_{\iota}}},\dfrac{1}{\ell_{r}}\uplambda_{\mu},\sqrt{\frac{\rho_{r}}{\mu_{r}}}\hskip 1.13809pt\textsf{c}_{\mu}\big{)}, ι=1,2,3,\displaystyle\!\!\!\!\!\!\iota=1,2,3,
(𝗁,𝖿,𝗏β)(h,f,vβ):=(1r𝗁,rρrμr𝖿,ρrμr𝗏β),\displaystyle({\sf h},{\sf f},{\sf{v}^{\upbeta}})~{}\rightarrow~{}(h,f,v^{\upbeta}):=\big{(}\dfrac{1}{\ell_{r}}{\sf h},\ell_{r}\sqrt{\frac{\rho_{r}}{\mu_{r}}}\hskip 1.13809pt{\sf f},\sqrt{\frac{\rho_{r}}{\mu_{r}}}\hskip 1.13809pt{\sf{v}^{\upbeta}}\big{)}, β=1,2,\displaystyle\!\!\!\!\!\!\upbeta=1,2,

where 𝗁=1.5{\sf h}=1.5 mm and 𝖿{\sf f} respectively indicate the specimen’s thickness and cyclic frequency of wave motion. Table 5 (resp. Table 6) details the normal values for the first (resp. second) of (10). The normal thickness and center frequencies are as follows,

{fc1,fc2,fc3}={0.33,0.16,0.59},h=0.15.\{{f_{c_{1}}},{f_{c_{2}}},{f_{c_{3}}}\}=\{0.33,0.16,0.59\},\quad h=0.15. (11)
Table 5: Properties of the aluminum, titanium, steel and brass sheets as per the manufacturer. Here, χμ:=Eμ/ρμ\chi_{\mu}\colon\!\!\!={E_{\mu}}/{\rho_{\mu}}.
physical μ\mu Al​ Ti​ St​ Br​
Eμ\textsf{E}_{\mu} [GPA] 68.9​ 105​ 199.95​ 110​
quantity ρμ\uprho_{\mu}[kg/m3] 2700​ 4510​ 7850​ 8530​
νμ\nu_{\mu} 0.33​ 0.34​ 0.29​ 0.31​
normal EμE_{\mu} 1​ 1.52​ 2.90​ 1.60​
value ρμ\rho_{\mu} 1​ 1.67​ 2.91​ 3.16​
χμ\chi_{\mu} 1​ 0.91​ 1​ 0.51​
Table 6: Phase velocity cμ\textsf{c}_{\mu} and wavelength λμ\uplambda_{\mu} in μ={Al, Ti, St, Br}\mu=\{\text{Al, Ti, St, Br}\} at {𝖿𝖼𝟣,𝖿𝖼𝟤,𝖿𝖼𝟥}={165,80,300}\{{\sf f_{c_{1}}},{\sf f_{c_{2}}},{\sf f_{c_{3}}}\}=\{165,80,300\} kHz as per Fig. 19, and their normalized counterparts according to (10).

physical quantity μ\mu Al​ Ti​ St​ Br​ λμ(𝖿𝖼𝟣)\uplambda_{\mu}({\sf f_{c_{1}\!}}) [cm] 11 - - - cμ(𝖿𝖼𝟣)\textsf{c}_{\mu}({\sf f_{c_{1}\!}})[m/s] 1610.41610.4 - - - λμ(𝖿𝖼𝟤)\uplambda_{\mu}({\sf f_{c_{2}\!}}) [cm] - 1.41.4 1.41.4 1.171.17 cμ(𝖿𝖼𝟤)\textsf{c}_{\mu}({\sf f_{c_{2}\!}})[m/s] - 11401140 11261126 936936 λμ(𝖿𝖼𝟥)\uplambda_{\mu}({\sf f_{c_{3}\!}}) [cm] - 0.650.65 0.640.64 0.50.5 cμ(𝖿𝖼𝟥)\textsf{c}_{\mu}({\sf f_{c_{3}\!}})[m/s] - 1960.81960.8 19291929 1501.61501.6

normal value μ\mu Al​ Ti​ St​ Br​ λμ(fc1)\lambda_{\mu}({f_{c_{1}\!}}) 11 - - - cμ(fc1){c}_{\mu}({f_{c_{1}\!}}) 0.320.32 - - - λμ(fc2)\lambda_{\mu}({f_{c_{2}\!}}) - 1.41.4 1.41.4 1.171.17 cμ(fc2){c}_{\mu}({f_{c_{2}\!}}) - 0.230.23 0.220.22 0.190.19 λμ(fc3)\lambda_{\mu}({f_{c_{3}\!}}) - 0.650.65 0.640.64 0.50.5 cμ(fc3){c}_{\mu}({f_{c_{3}\!}}) - 0.390.39 0.380.38 0.30.3

4.3 Governing equation

In light of (11) and Table 6, observe that in all tests the wavelength-to-thickness ratio λμh[3.33   9.33]\frac{\lambda_{\mu}}{h}\in[3.33\,\,\,9.33], μ={Al, Ti, St, Br}\mu=\{\text{Al, Ti, St, Br}\}. Therefore, one may invoke the equation governing flexural waves in thin plates [53] to approximate the physics of measured data. In this framework, (1) may be recast as

Λ=Λβ:=χβh312(1νβ2)4h(2πf)2,\displaystyle\!\!\Lambda~{}=~{}\Lambda_{\upbeta}~{}\colon\!\!\!=~{}\frac{\chi_{\upbeta}h^{3}}{12(1-\nu_{\upbeta}^{2})}\nabla^{4}\,-\,h(2\pi{f})^{2}, χβ:=Eβρβ,β= 1,2,\displaystyle\chi_{\upbeta}\,\colon\!\!\!=\,\frac{E_{\upbeta}}{\rho_{\upbeta}},\,\upbeta\,=\,1,2, (12)
u^=vβ(𝝃,f;𝝉),ϑ=χβ(𝝃,f),\displaystyle\!\!\hat{\text{{u}}}~{}=~{}v^{\upbeta}(\boldsymbol{\xi},f;{\boldsymbol{\tau}}),\quad\boldsymbol{\vartheta}~{}=~{}\chi_{\upbeta}(\boldsymbol{\xi},f), 𝝃Sobs,𝝉Sinc,f[0.8   1.2]fcι,ι=1,2,3,\displaystyle\boldsymbol{\xi}\in S^{\text{obs}}\!,\,{\boldsymbol{\tau}}\in S^{\text{inc}},\,f\in[0.8\,\,\,1.2]f_{c_{\iota}},\,\iota=1,2,3,

where ρβ,Eβ,νβ\rho_{\upbeta},E_{\upbeta},\nu_{\upbeta} respectively denote the normal density, Young’s modulus, and Poisson’s ratio in Πβexp\Uppi_{\upbeta}^{{}^{\text{exp}}}\!, β= 1,2\upbeta\,=\,1,2, and 𝝉{\boldsymbol{\tau}} indicates the source location. Note that νβ0.32\nu_{\upbeta}\sim 0.32 according to Table 5 and Λ\Lambda, related to 1νβ21-\nu_{\upbeta}^{2}, shows little sensitivity to small variations in the Poisson’s ratio. Thus, in what follows, νβ\nu_{\upbeta} is treated as a known parameter. Provided vβ(𝝃,f;𝝉)v^{\upbeta}(\boldsymbol{\xi},f;{\boldsymbol{\tau}}), the objective is to reconstruct χβ(𝝃,f)\chi_{\upbeta}(\boldsymbol{\xi},f).

4.4 Direct inversion

Following the reconstruction procedure of Section 3.2, the distribution of χβ\chi_{\upbeta} in 𝒢β\mathscr{G}_{\upbeta}, β=1,2\upbeta=1,2, is obtained at specific frequencies. In this vein, the positive-definite MLP networks χβ=𝒩χβ(𝝃,ω)\chi^{\star}_{\upbeta}=\mathscr{N}_{\chi_{\upbeta}}(\boldsymbol{\xi},\omega) and α=𝒩α(𝝃,ω)\alpha^{\star}=\mathscr{N}_{\alpha}(\boldsymbol{\xi},\omega) comprised of three hidden layers of respectively 2020, 4040, and 2020 neurons are constructed according to Fig. 1. In all MLP trainings of this section, each epoch makes use of the full dataset and the learning rate is 0.0050.005.

When β=1\upbeta=1, the inversion is conducted at f1=0.336f_{1}=0.336. SincS^{\text{inc}} is sampled at one point i.e., the piezoelectric transducer remains fixed during the test on Al plate, and thus, Nτ=1N_{\tau}=1, while a concentric 60×6060\!\times\!60 subset of collocation points sampling SobsS^{\text{obs}} is deployed for training. In this setting, the input to χ1\chi^{\star}_{1} and α\alpha^{\star} is of size NξNτ×Nω=3600×1N_{\xi}N_{\tau}\times N_{\omega}=3600\times 1, and their real-valued outputs are of the same size. The results are shown in Fig. 16. When β=2\upbeta=2, the direct inversion is conducted at f2=0.17f_{2}=0.17 and f3=0.61f_{3}=0.61. For the low-frequency reconstruction, SincS^{\text{inc}} is sampled at one point, while a 40×12040\!\times\!120 subset of scan points in 𝒢2\mathscr{G}_{2} is used for training so that the input/output size for χ2\chi^{\star}_{2} and α\alpha^{\star} is 4600×14600\!\times\!1. The recovered fields and associated normal error are provided in Fig. 17. Table 7 enlists the true values as well as mean and standard deviation of the reconstructed distributions χβ\chi^{\star}_{\upbeta} in Πβexp\Uppi_{\upbeta}^{{}^{\text{exp}}}\!, β=1,2\upbeta=1,2, according to Figs. 16 and 17. For the high-frequency reconstruction, when β=2\upbeta=2, SincS^{\text{inc}} is sampled at three points i.e., experiments are performed for three distinct positions of the piezoelectric transducer, while the same subset of scan points is used for training. In this case, the input to χ2\chi^{\star}_{2} and α\alpha^{\star} is 13800×113800\!\times\!1, while their output is of dimension 4600×14600\!\times\!1. The high-frequency reconstruction results are illustrated in Fig. 18, and the affiliated means and standard deviations are provided in Table 8. It should be mentioned that the computed normal errors in Figs. 1617, and 18 are with respect to the verified values of Section 4.4.1. Note that the recovered α\alpha^{\star}s from laboratory test data are much smoother than the ones reconstructed from synthetic data in Section 3.2. This could be attributed to the scaler nature of (12) with a single unknown parameter – as opposed to the vector equations governing the in-plane wave motion with two unknown parameters. More specifically, here, α\alpha^{\star} controls the weights and biases of a single network χβ\chi^{\star}_{\upbeta}, while in Section 3.2, α\alpha^{\star} simultaneously controls the parameters of two separate networks μ\mu^{\star} and λ\lambda^{\star}. A comparative analysis of Figs. 17 and 18 reveals that (a) enriching the waveform data by increasing the number of sources remarkably decrease the reconstruction error, (b) the regularization parameter α\alpha in (3) is truly distributed in nature as the magnitude of the recovered α\alpha^{\star} in brass is ten times greater than that of titanium and steel which is due to the difference in the level of noise in measurements related to distinct material surfaces, and (c) the recovered field χ2\chi^{\star}_{2} – which according to (12) is a material property E2/ρ2{E_{2}}/{\rho_{2}}, demonstrates a significant dependence to the reconstruction frequency. The latter calls for proper verification of the results which is the subject of Section 4.4.1.

4.4.1 Verification

To shine some light on the nature discrepancies between the low- and high- frequency reconstructions in

Refer to caption
Figure 16: Direct inversion of the PDE parameter χ1\chi_{1} in Π1exp\Uppi_{1}^{{}^{\text{exp}}}\! using test data from a single source at frequency f1=0.336f_{1}=0.336: (a) MLP-predicted distribution χ1(𝝃,f1)\chi_{1}(\boldsymbol{\xi},f_{1}) in 𝝃𝒢1\boldsymbol{\xi}\in\mathscr{G}_{1}, (b) reconstruction error (8) with respect to the true value χ1=χAl=1\chi_{1}=\chi_{\text{Al}}=1, (c) MLP-recovered distribution of the regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e} in log scale.
Refer to caption
Figure 17: Direct inversion of the PDE parameter χ2\chi_{2} in Π2exp\Uppi_{2}^{{}^{\text{exp}}}\! using test data from a single source at frequency f2=0.17f_{2}=0.17: (a) MLP-predicted distribution χ2(𝝃,f2)\chi_{2}(\boldsymbol{\xi},f_{2}) in 𝝃𝒢2\boldsymbol{\xi}\in\mathscr{G}_{2}, (b) reconstruction error (8) with respect to the true value χ2{χTi,χSt,χBr}={0.91,1,0.51}\chi_{2}\in\{\chi_{\text{Ti}},\chi_{\text{St}},\chi_{\text{Br}}\}=\{0.91,1,0.51\} as per Table 5, (c) MLP-recovered distribution of the regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e} in log scale.
Table 7: Mean and standard deviation of the reconstructed distributions in Figs. 16 and 17 via the direct inversion of single-source test data.
β\upbeta 11 2Ti2_{\text{Ti}} 2St2_{\text{St}} 2Br2_{\text{Br}}
χβ\chi_{\upbeta} 11 0.910.91 11 0.510.51
χβΠβexp\hskip 1.13809pt\langle\hskip 0.56905pt\chi_{\upbeta}^{\star}\rangle_{\Uppi_{\upbeta}^{{}^{\text{exp}}}\!} 1.0411.041 0.8720.872 0.9780.978 0.4430.443
σ(χβ|Πβexp)\upsigma(\chi_{\upbeta}^{\star}|_{\Uppi_{\upbeta}^{{}^{\text{exp}}}\!}) ​​0.0170.017​​ ​​0.0440.044​​ ​​0.0600.060​​ ​​0.0520.052​​
Refer to caption
Figure 18: Direct inversion of the PDE parameter χ2\chi_{2} in Π2exp\Uppi_{2}^{{}^{\text{exp}}}\! using test data from three source locations at frequency f3=0.61f_{3}=0.61: (a) MLP-predicted distribution χ2(𝝃,f3)\chi_{2}(\boldsymbol{\xi},f_{3}) in 𝝃𝒢2\boldsymbol{\xi}\in\mathscr{G}_{2}, (b) reconstruction error (8) with respect to the related estimates {0.57,0.59,0.24}\{0.57,0.59,0.24\} as per Fig. 20, (c) MLP-recovered distribution of the regularization parameter α\alpha^{\star}, and (d) loss function ε\mathscr{L}_{\varepsilon} vs. the number of epochs NeN_{e} in log scale.
Table 8: Mean and standard deviation of the reconstructed distributions in Fig. 18 via the direct inversion applied to high-frequency test data from three distinct sources.
β\upbeta 2Ti2_{\text{Ti}} 2St2_{\text{St}} 2Br2_{\text{Br}}
χβ\chi^{\prime}_{\upbeta} 0.570.57 0.590.59 0.240.24
χβΠβexp\hskip 1.13809pt\langle\hskip 0.56905pt\chi_{\upbeta}^{\star}\rangle_{\Uppi_{\upbeta}^{{}^{\text{exp}}}\!} ​​0.5850.585​​ ​​0.6060.606​​ ​​0.2270.227​​
σ(χβ|Πβexp)\upsigma(\chi_{\upbeta}^{\star}|_{\Uppi_{\upbeta}^{{}^{\text{exp}}}\!}) ​​0.0150.015​​ ​​0.0290.029​​ ​​0.0160.016​​
Refer to caption
Figure 19: Experimental vs. theoretical dispersion curves 𝖿(λμ1){\sf f}(\uplambda_{\mu}^{-1}) for μ={Al, Ti, St, Br}\mu=\{\text{Al, Ti, St, Br}\}. Analytical curves (solid lines) are computed from (13) using the pertinent properties in Table 5.
Refer to caption
Figure 20: Discrepancy in the balance law (12) at f3=0.61f_{3}=0.61: (a) elastic force field 𝔗μ1\mathfrak{T}_{\mu}^{1}, μ={Ti, St, Br}\mu=\{\text{Ti, St, Br}\}, according to (14) with adjusted coefficients {χTi,χSt,χBr}={0.57,0.59,0.24}\{\chi_{\text{Ti}},\chi_{\text{St}},\chi_{\text{Br}}\}=\{0.57,0.59,0.24\}, (b) the inertia field 𝔗μ2\mathfrak{T}_{\mu}^{2}, and (c) normal discrepancy 𝔇μ\mathfrak{D}_{\mu}.

Figs. 17 and 18, a set of secondary tests are performed to obtain the dispersion curve for each component of the test setup. For this purpose, antiplane shear waves of form (9) are induced at 𝖿𝖼𝗃=50j{\sf f_{c_{j}}}\!=50j kHz, j=1,2,,7j=1,2,\ldots,7, in 6060 cm ×\!\times\! 6060 cm cuts of aluminum, titanium, steel, and brass sheets used in the primary tests of Fig. 15. In each experiment, the piezoelectric transducer is placed in the middle of specimen (far from the external boundary), and the out-of-plane wave motion is captured in the immediate vicinity of the transducer along a straight line of length 88 cm sampled at 400400 scan points. The Fourier-transformed signals in time-space furnish the dispersion relations of Fig. 19. In parallel, the theoretical dispersion curves affiliated with (12) are computed according to

𝖿=2π(λμ)2χμ𝗁212(1νμ2),χμ=𝖤μρμ,μ={Al, Ti, St, Br},{\sf f}\!~{}=~{}2\pi(\uplambda_{\mu})^{-2}\sqrt{\frac{\chi_{\mu}{\sf h}^{2}}{12(1-\nu_{\mu}^{2})}},\quad\chi_{\mu}~{}=~{}\frac{{\sf E}_{\mu}}{\uprho_{\mu}},\quad\mu=\{\text{Al, Ti, St, Br}\}, (13)

using the values of Table 5 for χμ\chi_{\mu} and νμ\nu_{\mu} and 𝗁=1.5{\sf h}=1.5mm. A comparison between the experimental and theoretical dispersion curves 𝖿(λμ1){\sf f}(\uplambda_{\mu}^{-1}) in Fig. 19 verifies the theory and the values of Table 5 for χμ\chi_{\mu} in the low-frequency regime of wave motion. This is also in agreement with the direct inversion results of Figs. 16 and 17. Moreover, Fig. 19 suggests that at approximately 𝖿μ={170,200,120,110}{\sf f}_{\mu}=\{{170,200,120,110}\} kHz for μ={Al, Ti, St, Br}\mu=\{\text{Al, Ti, St, Br}\} the governing PDE (12) with physical coefficients fails to predict the experimental results which may provide an insight regarding the high-frequency reconstruction results in Fig. 18. Further investigation of the balance law (12), as illustrated in Fig. 20, shows that the test data at 312312 kHz satisfy – with less than 1020%10-20\% discrepancy depending on the material – a PDE of form (12) with modified coefficients. More specifically, Fig. 20 demonstrates the achievable balance between the elastic force distribution 𝔗μ1\mathfrak{T}_{\mu}^{1} and inertia field 𝔗μ2\mathfrak{T}_{\mu}^{2} in (12) by directly adjusting the PDE parameter χ2\chi^{\prime}_{2} to minimize the discrepancy 𝔇μ\mathfrak{D}_{\mu} according to

𝔗μ1:=χ2h312(1ν22)4v2,𝔗μ2:=h(2πf)2v2,𝔇μ:=|𝔗μ1𝔗μ2|max|𝔗μ2|.\displaystyle\!\!\mathfrak{T}_{\mu}^{1}~{}\colon\!\!\!=~{}\frac{\chi^{\prime}_{2}h^{3}}{12(1-\nu_{2}^{2})}\nabla^{4}v^{2},\quad\mathfrak{T}_{\mu}^{2}~{}\colon\!\!\!=~{}h(2\pi{f})^{2}v^{2},\quad\mathfrak{D}_{\mu}~{}\colon\!\!\!=~{}\frac{|\mathfrak{T}_{\mu}^{1}-\mathfrak{T}_{\mu}^{2}|}{\max|\mathfrak{T}_{\mu}^{2}|}. (14)

With reference to Table 8, the recovered coefficients χ2\chi^{\prime}_{2} at f=f3=0.61f=f_{3}=0.61 verify the direct inversion results of Fig. 18. This implies that the direct inversion (or PINNs) may lead to non-physical reconstructions in order to attain the best fit for the data to the “perceived”” underlying physics. Thus, it is imperative to establish the range of validity of the prescribed physical principles in data-driven modeling. Here, the physics of the system at f3f_{3} is in transition, yet close enough to the leading-order approximation (12) that the discrepancy is less than 20%20\%. It is unclear, however, if this equation with non-physical coefficients may be used as a predictive tool. It would be interesting to further investigate the results through the prism of higher-order continuum theories and a set of independent experiments for validation which could be the subject of a future study.

4.5 Physics-informed neural networks

Following Section 3.3, PINNs are built and trained using experimental test data of Section 4.4. The MLP network v1=v1(𝝃,f,𝒙|γ,χ1){v^{1}}^{\star}={v^{1}}^{\star}(\boldsymbol{\xi},f,\boldsymbol{x}\hskip 1.13809pt|\hskip 1.13809pt\gamma,\chi_{1}^{\star}) with six hidden layers of respectively 4040, 4040, 120120, 8080, 4040, and 4040 neurons is constructed to map the out-of-plane velocity field v1{v^{1}} (in Π1exp\Uppi_{1}^{{}^{\text{exp}}}) related to a single transducer location 𝒙1\boldsymbol{x}_{1} and frequency f1=0.336f_{1}=0.336. The PDE parameter χ1\chi_{1} is defined as the unknown scaler parameter of the network, and following the argument of Section 3.3, the Lagrange multiplier γ\gamma is specified as a nonadaptive scaler weight of magnitude 1h(2πf1)2=1.5\frac{1}{h(2\pi f_{1})^{2}}=1.5. The input/output dimension for v1{v^{1}}^{\star} is Nξ×Nω×Nτ=3600×1×1N_{\xi}\!\times\!N_{\omega}\!\times\!N_{\tau}=3600\!\times\!1\!\times\!1, and each epoch makes use of the full dataset for training and the learning rate is 0.0050.005. Keep in mind that the objective here is to (a) construct a surrogate map for v1{v^{1}}, and (b) identify χ1\chi_{1}^{\star}.

Fig. 21 demonstrates (a) the accuracy of PINN-estimated field v1{v^{1}}^{\star} compared to the test data v1v^{1}, (b) performance of automatic differentiation in capturing the fourth-order field derivatives e.g., v1,1111{{v}^{1}}^{\star}_{\!\!,1111} that appear in the governing PDE (12), and (c) the evolution of parameter χ1\chi_{1}^{\star}. The comparison in (b) is with respect to the spectral derivates of test data according to Section 2.3.2. It is no surprise that the automatic differentiation incurs greater errors in estimating the higher order derivatives involved in the antiplane wave motion compared to the second-order derivatives of Section 3.3.

In addition, the PINN v2=v2(𝝃,f,𝒙|γ,χ2){v^{2}}^{\star}={v^{2}}^{\star}(\boldsymbol{\xi},f,\boldsymbol{x}\hskip 1.13809pt|\hskip 1.13809pt\gamma,\chi_{2}^{\star}) with seven hidden layers of respectively 4040, 4040, 120120, 120120, 8080, 4040, and 4040 neurons is created to reconstruct (i) particle velocity field v2{v^{2}} in the layered specimen Π2exp\Uppi_{2}^{{}^{\text{exp}}}, and (ii) distribution of the PDE parameter χ2\chi_{2} in the sampling area. The latter is defined as an unknown parameter of the network with dimension 40×12040\!\times\!120, and the scaler weight γ\gamma is set to 1h(2πf2)2=5.84\frac{1}{h(2\pi f_{2})^{2}}=5.84 for the low-frequency reconstruction. In this setting, the input/output dimension for v2{v^{2}}^{\star} reads 4800×1×14800\!\times\!1\!\times\!1. Fig. 22 provides a comparative analysis between the experimental and PINN-predicted maps of velocity and PDE parameter. The associated statistics are provided in Table 9. It is evident from the waveform in Fig. 22 (a) that the most pronounced errors in Fig. 22 (d) occur at the loci of vanishing particle velocity. Similar to Section 3.2, this could be potentially addressed by enriching the test data.

5 Conclusions

The ML-based direct inversion and physics-informed neural networks are investigated for full-field ultrasonic characterization of layered components. Direct inversion makes use of signal processing tools to directly compute the field derivatives from dense datasets furnished by laser-based ultrasonic experiments. This allows for a simplified and controlled learning process that specifically recovers the sought-for physical fields through minimizing a single-objective loss function. PINNs are by design more versatile and particularly advantageous with limited test data where waveform completion is desired (or required) for mechanical characterization. PINNs multi-objective learning from ultrasonic data may be more complex but can be accomplished via carefully gauged loss functions.

In direct inversion, Tikhonov regularization is critical for stable reconstruction of distributed PDE parameters from limited or multi-fidelity experimental data. In this vein, deep learning offers a unique opportunity to simultaneously recover the regularization parameter as an auxiliary field which proved to be particularly insightful in inversion of experimental data.

In training PINNs, two strategies were remarkably helpful: (1) identifying the reference length scale by the dominant wavelength in an effort to control the norm of spatial derivatives – which turned out to be crucial in the case of flexural waves in thin plates with the higher order PDE, and (2) estimating the Lagrange multiplier by taking advantage of the inertia term in the governing PDEs.

Laboratory implementations at multiple frequencies exposed that verification and validation are indispensable for predictive data-driven modeling. More specifically, both direct inversion and PINNs recover the unknown “physical” quantities that best fit the data to specific equations (with often unspecified range of validity). This may lead to mathematically decent but physically incompatible reconstructions especially when the perceived physical laws are near their limits such that the discrepancy in capturing the actual physics is significant. In which case, the inversion algorithms try to compensate for this discrepancy by adjusting the PDE parameters which leads to non-physical reconstructions. Thus, it is paramount to conduct complementary experiments to (a) establish the applicability of prescribed PDEs, and (b) validate the predictive capabilities of the reconstructed models.

Refer to caption
Figure 21: PINN vs. experimental maps of particle velocity and its derivatives in Π1exp\Uppi_{1}^{{}^{\text{exp}}}: (a) PINN estimates for {v1,v1,1111,v1,2222,v1,1122}\{{v^{1}}^{\star}\!,{v^{1}}^{\star}_{\!\!\!,1111},{v^{1}}^{\star}_{\!\!\!,2222},{v^{1}}^{\star}_{\!\!\!,1122}\} wherein the derivatives are obtained by automatic differentiation, (b) normalized LDV-captured particle velocity field v1v^{1} and its corresponding spectral derivatives, (c) normal misfit 8 between (a) and (b), (d) PINN-reconstructed PDE parameter χ1\chi^{\star}_{1} vs. the number of epochs NeN_{e}, and (e) total loss ϖ\mathscr{L}_{\varpi} and its components (the PDE residue and data misfit) vs. NeN_{e} in log scale.
Refer to caption
Figure 22: Low-frequency PINN reconstruction in Π2exp\Uppi_{2}^{{}^{\text{exp}}}\! using test data from a single source at f2=0.17f_{2}=0.17: (a) PINN-predicted distribution of particle velocity v2{v^{2}}^{\star}, (b) normalized LDV-captured particle velocity v2v^{2}, (c) normal misfit between (a) and (b), (d) PINN-predicted distribution of the PDE parameter χ2\chi^{\star}_{2}, and (e) total loss ϖ\mathscr{L}_{\varpi} and its components (the PDE residue and data misfit) vs. the number of epochs NeN_{e} in log scale.
Table 9: Mean and standard deviation of the PINN-reconstructed distributions in Fig. 22 from a single-source, low-frequency test data.
β\upbeta 2Ti2_{\text{Ti}} 2St2_{\text{St}} 2Br2_{\text{Br}}
χβ\chi_{\upbeta} 0.910.91 11 0.510.51
χβΠβexp\hskip 1.13809pt\langle\hskip 0.56905pt\chi_{\upbeta}^{\star}\rangle_{\Uppi_{\upbeta}^{{}^{\text{exp}}}\!} 0.7900.790 0.8900.890 0.4140.414
σ(χβ|Πβexp)\upsigma(\chi_{\upbeta}^{\star}|_{\Uppi_{\upbeta}^{{}^{\text{exp}}}\!}) ​​0.2140.214​​ ​​0.3560.356​​ ​​0.1340.134​​

Authors’ contributions

Y.X. investigation, methodology, data curation, software, visualization, writing – original draft; F.P. conceptualization, methodology, funding acquisition, supervision, writing – original draft; J.S. experimental data curation; C.W. experimental data curation.

Acknowledgments

This study was funded by the National Science Foundation (Grant No. 1944812) and the University of Colorado Boulder through FP’s startup. This work utilized resources from the University of Colorado Boulder Research Computing Group, which is supported by the National Science Foundation (awards ACI-1532235 and ACI-1532236), the University of Colorado Boulder, and Colorado State University. Special thanks are due to Kevish Napal for facilitating the use of FreeFem++ code developed as part of [49] for elastodynamic simulations.

References

  • [1] X. Liang, M. Orescanin, K. S. Toohey, M. F. Insana, S. A. Boppart, Acoustomotive optical coherence elastography for measuring material mechanical properties, Optics letters 34 (19) (2009) 2894–2896.
  • [2] G. Bal, C. Bellis, S. Imperiale, F. Monard, Reconstruction of constitutive parameters in isotropic linear elasticity from noisy full-field measurements, Inverse problems 30 (12) (2014) 125004.
  • [3] B. S. Garra, Elastography: history, principles, and technique comparison, Abdominal imaging 40 (4) (2015) 680–697.
  • [4] C. Bellis, H. Moulinec, A full-field image conversion method for the inverse conductivity problem with internal measurements, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 472 (2187) (2016) 20150488.
  • [5] H. Wei, T. Mukherjee, W. Zhang, J. Zuback, G. Knapp, A. De, T. DebRoy, Mechanistic models for additive manufacturing of metallic components, Progress in Materials Science 116 (2021) 100703.
  • [6] C.-T. Chen, G. X. Gu, Learning hidden elasticity with deep neural networks, Proceedings of the National Academy of Sciences 118 (31) (2021) e2102721118.
  • [7] H. You, Q. Zhang, C. J. Ross, C.-H. Lee, M.-C. Hsu, Y. Yu, A physics-guided neural operator learning approach to model biological tissues from digital image correlation measurements, arXiv preprint arXiv:2204.00205.
  • [8] C. M. Bishop, N. M. Nasrabadi, Pattern recognition and machine learning, Vol. 4, Springer, 2006.
  • [9] Y. LeCun, Y. Bengio, G. Hinton, Deep learning, nature 521 (7553) (2015) 436–444.
  • [10] S. Cuomo, V. S. Di Cola, F. Giampaolo, G. Rozza, M. Raissi, F. Piccialli, Scientific machine learning through physics-informed neural networks: Where we are and what’s next, arXiv preprint arXiv:2201.05624.
  • [11] S. Wang, H. Wang, P. Perdikaris, Improved architectures and training algorithms for deep operator networks, Journal of Scientific Computing 92 (2) (2022) 1–42.
  • [12] L. McClenny, U. Braga-Neto, Self-adaptive physics-informed neural networks using a soft attention mechanism, arXiv preprint arXiv:2009.04544.
  • [13] Z. Chen, V. Badrinarayanan, C.-Y. Lee, A. Rabinovich, Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks, in: International conference on machine learning, PMLR, 2018, pp. 794–803.
  • [14] M. Raissi, P. Perdikaris, G. E. Karniadakis, Physics informed deep learning (part i): Data-driven solutions of nonlinear partial differential equations, arXiv preprint arXiv:1711.10561.
  • [15] M. Raissi, P. Perdikaris, G. E. Karniadakis, Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations, Journal of Computational physics 378 (2019) 686–707.
  • [16] E. Haghighat, M. Raissi, A. Moure, H. Gomez, R. Juanes, A physics-informed deep learning framework for inversion and surrogate modeling in solid mechanics, Computer Methods in Applied Mechanics and Engineering 379 (2021) 113741.
  • [17] A. Henkes, H. Wessels, R. Mahnken, Physics informed neural networks for continuum micromechanics, Computer Methods in Applied Mechanics and Engineering 393 (2022) 114790.
  • [18] R. Muthupillai, D. Lomas, P. Rossman, J. F. Greenleaf, A. Manduca, R. L. Ehman, Magnetic resonance elastography by direct visualization of propagating acoustic strain waves, science 269 (5232) (1995) 1854–1857.
  • [19] P. E. Barbone, N. H. Gokhale, Elastic modulus imaging: on the uniqueness and nonuniqueness of the elastography inverse problem in two dimensions, Inverse problems 20 (1) (2004) 283.
  • [20] O. A. Babaniyi, A. A. Oberai, P. E. Barbone, Direct error in constitutive equation formulation for plane stress inverse elasticity problem, Computer methods in applied mechanics and engineering 314 (2017) 3–18.
  • [21] A. N. Tikhonov, A. Goncharsky, V. Stepanov, A. G. Yagola, Numerical methods for the solution of ill-posed problems, Vol. 328, Springer Science & Business Media, 1995.
  • [22] A. Kirsch, et al., An introduction to the mathematical theory of inverse problems, Vol. 120, Springer, 2011.
  • [23] On the convergence of physics informed neural networks for linear second-order elliptic and parabolic type pdes, Communications in Computational Physics 28 (5) (2020) 2042–2074.
  • [24] S. Wang, X. Yu, P. Perdikaris, When and why pinns fail to train: A neural tangent kernel perspective, Journal of Computational Physics 449 (2022) 110768.
  • [25] Z. Xiang, W. Peng, X. Liu, W. Yao, Self-adaptive loss balanced physics-informed neural networks, Neurocomputing 496 (2022) 11–34.
  • [26] R. Bischof, M. Kraus, Multi-objective loss balancing for physics-informed deep learning, arXiv preprint arXiv:2110.09813.
  • [27] H. Son, S. W. Cho, H. J. Hwang, Al-pinns: Augmented lagrangian relaxation method for physics-informed neural networks, arXiv preprint arXiv:2205.01059.
  • [28] S. Zeng, Z. Zhang, Q. Zou, Adaptive deep neural networks methods for high-dimensional partial differential equations, Journal of Computational Physics 463 (2022) 111232.
  • [29] J. Yu, L. Lu, X. Meng, G. E. Karniadakis, Gradient-enhanced physics-informed neural networks for forward and inverse pde problems, Computer Methods in Applied Mechanics and Engineering 393 (2022) 114823.
  • [30] S. Wang, Y. Teng, P. Perdikaris, Understanding and mitigating gradient flow pathologies in physics-informed neural networks, SIAM Journal on Scientific Computing 43 (5) (2021) A3055–A3081.
  • [31] A. D. Jagtap, K. Kawaguchi, G. Em Karniadakis, Locally adaptive activation functions with slope recovery for deep and physics-informed neural networks, Proceedings of the Royal Society A 476 (2239) (2020) 20200334.
  • [32] Y. Kim, Y. Choi, D. Widemann, T. Zohdi, A fast and accurate physics-informed neural network reduced order model with shallow masked autoencoder, Journal of Computational Physics 451 (2022) 110841.
  • [33] G. I. Barenblatt, Scaling (Cambridge texts in applied mathematics), Cambridge University Press, Cambridge, UK, 2003.
  • [34] K. Hornik, Approximation capabilities of multilayer feedforward networks, Neural networks 4 (2) (1991) 251–257.
  • [35] Y. Chen, L. Dal Negro, Physics-informed neural networks for imaging and parameter retrieval of photonic nanostructures from near-field data, APL Photonics 7 (1) (2022) 010802.
  • [36] F. Pourahmadian, B. B. Guzina, On the elastic anatomy of heterogeneous fractures in rock, International Journal of Rock Mechanics and Mining Sciences 106 (2018) 259 – 268.
  • [37] X. Liu, J. Song, F. Pourahmadian, H. Haddar, Time-vs. frequency-domain inverse elastic scattering: Theory and experiment, arXiv preprint arXiv:2209.07006.
  • [38] F. Pourahmadian, B. B. Guzina, H. Haddar, Generalized linear sampling method for elastic-wave sensing of heterogeneous fractures, Inverse Problems 33 (5) (2017) 055007.
  • [39] F. Cakoni, D. Colton, H. Haddar, Inverse Scattering Theory and Transmission Eigenvalues, SIAM, 2016.
  • [40] V. A. Morozov, Methods for solving incorrectly posed problems, Springer Science & Business Media, 2012.
  • [41] R. Kress, Linear integral equation, Springer, Berlin, 1999.
  • [42] A. Paszke, S. Gross, S. Chintala, G. Chanan, E. Yang, Z. DeVito, Z. Lin, A. Desmaison, L. Antiga, A. Lerer, Automatic differentiation in pytorch.
  • [43] H. Brezis, Functional analysis, Sobolev spaces and partial differential equations, Springer Science & Business Media, 2010.
  • [44] T. Ha-Duong, On retarded potential boundary integral equations and their discretization, in: Topics in computational wave propagation, Springer, 2003, pp. 301–336.
  • [45] R. T. Rockafellar, Lagrange multipliers and optimality, SIAM review 35 (2) (1993) 183–238.
  • [46] H. Everett III, Generalized lagrange multiplier method for solving problems of optimum allocation of resources, Operations research 11 (3) (1963) 399–417.
  • [47] D. Liu, Y. Wang, A dual-dimer method for training physics-constrained neural networks with minimax architecture, Neural Networks 136 (2021) 112–125.
  • [48] F. Hecht, New development in freefem++, Journal of Numerical Mathematics 20 (3-4) (2012) 251–265.
    URL https://freefem.org/
  • [49] F. Pourahmadian, K. Napal, Poroelastic near-field inverse scattering, Journal of Computational Physics 455 (2022) 111005.
  • [50] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Kopf, E. Yang, Z. DeVito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, S. Chintala, Pytorch: An imperative style, high-performance deep learning library, Advances in neural information processing systems 32.
  • [51] D. P. Kingma, J. Ba, Adam: A method for stochastic optimization, arXiv preprint arXiv:1412.6980.
  • [52] Pytorch implementation of physics-informed neural networks, https://github.com/jayroxis/PINNs (2022).
  • [53] K. F. Graff, Wave motion in elastic solids, Courier Corporation, 2012.