Professional Documents
Culture Documents
Abstract
In this work, we present the physics-informed neural network (PINN) model applied particularly to dynamic problems in solid
mechanics. We focus on forward and inverse problems. Particularly, we show how a PINN model can be used efficiently for
material identification in a dynamic setting. In this work, we assume linear continuum elasticity. We show results for two-
dimensional (2D) plane strain problem and then we proceed to apply the same techniques for a three-dimensional (3D) problem.
As for the training data we use the solution based on the finite element method. We rigorously show that PINN models are
accurate, robust and computationally efficient, especially as a surrogate model for material identification problems. Also, we
arXiv:2312.15175v1 [cs.NE] 23 Dec 2023
employ state-of-the-art techniques from the PINN literature which are an improvement to the vanilla implementation of PINN.
Based on our results, we believe that the framework we have developed can be readily adapted to computational platforms for
solving multiple dynamic problems in solid mechanics.
1. Introduction
Dynamic problems in solid mechanics are ever-present in industrial settings across multiple domains such as automotive, ship
building, aerospace to name a few. Traditionally, experiments or measurements are conducted to analyze these problems. In
the event that experiments are practically unfeasible to perform or in the case of repetitive need for it which is more often the
case, numerical methods had come to the forefront for obtaining solutions for these problems for many decades. For example
the finite element method (FEM) (Zienkiewicz and Taylor 2005) has been one of the most useful numerical techniques to solve
many problems in the area of solid mechanics. However, in many situations, for example, in material parameter identification
problems, numerical simulations incur high computational costs. Nevertheless, they are extensively used until the simulation time
is beyond an acceptable threshold. So, the practical question of whether such problems can be solved otherwise is a motivation to
seek newer ideas and methods. One such area of recent interest is the rapidly growing field of deep learning (LaCun et al. 2015)
which is based on the availability of large data. Deep learning has had great impact on fields such as computer vision (Elsayed
et al. 2018; Liu et al. 2022), autonomous driving (Grigorescu et al. 2019; Fujiyoshi et al. 2019), healthcare (R. et al. 2018; Esteva
et al. 2018) among many. There has also been quite some development on deep learning based research for solving problems in
mechanics (Brunton et al. 2020; Kutz 2017; Sun et al. 2021).
Of particular interest in this work is their use in problems based on solid mechanics. Although deep learning has seeing a
recent upsurge in applications due to the availability of advanced computing resources, it has been used as early as the late 1980s
for simple problems withing the context of building structural performance assessment (Adeli and Yeh 1989). Deep learning has
also been used multiple times for constitutive modeling in solids over the years (Ghaboussi et al. 1991; Mozaffar et al. 2019).
Constitutive modeling aims to construct a mathematical form of the relationship between the applied stress to the material and
the resulting strain. Many a time, for complex materials or assembly of multiple components, there are gaps in the constitutive
models and deep learning based on data addresses these shortcomings. Recent advances in deep learning are also used towards
structural health monitoring (Sun et al. 2021; Yoon et al. 2022).
Usually, deep learning approaches use a lot of data and they don’t account for underlying physical laws of the system under
study. So, they may be unfeasible to use if there is a dearth of data. One such method which circumvents this issue while
using deep learning is known as physics-informed neural network (PINN). It uses the information of governing equations of the
system along with the boundary and initial conditions for constructing the overall loss function which the neural network tries
to minimise. It has an effect of penalization of nonphysical solutions. It was introduced in the seminal work of (Raissi et al.
2019). Since then, there has been a plethora of research works on its development (Jagtap and Karniadakis 2020; Raissi et al.
2020; Jin et al. 2021; Wang et al. 2020; Hu et al. 2023) and applications (Lucor et al. 2022; Cai et al. 2021; Jagtap et al. 2022;
∗ Corresponding author
email address: gopinath.venkatesh2@in.bosch.com
Preprint
2
Kag et al. 2022). We are interested in this work in the use of PINN models for solving problems in solid mechanics. One of
the pioneering works on this topic is done by (Haghighat et al. 2021). They provide a framework for the use of PINN in the
context of forward, inverse and surrogate modeling. Since then, there has been significant applications of PINN in this area.
For example, in (Zhang et al. 2020) uses PINN to identify nonhomogeneous hyperelastic materials. (Anton and Wessels 2023)
perform material identification from both experimental data as well as full-field displacement data with noise. In the work of
(Wu et al. 2023), they apply PINN for identifying material properties in linear elasticity. Another important work was done by
(Haghighat et al. 2023) where they use PIN to characterize constitutive models for elastoplastic solids. (Oommen et al. 2022)
applied PINN to discover constitutive models in problems involving elastic and viscoplastic solids. Recently, (Roy et al. 2023)
also provide a PINN framework for solving linearly elastic problems such as an end-loaded cantilever beam and Kirchoff-Love
thin plate. Also, (Roy and Guha 2023) use PINN where they combine data and physics to solve problems in plasticity. PINN has
also been used to detect defects in materials in the work of (Zhang et al. 2022). Recently, use of PINN for complex geometries
is shown by Costabal et al. (2022) where they apply it on a hyperelastic problem. Aside from PINN, in the context of solid
mechanics, in (González et al. 2019), they use similar ideas of combining data and physical laws of thermodynamics to correct
existing constitutive models for problems in hyperelasticity.
While there has been reasonable amount of work on PINN applied to solid mechanics, one of the questions which still persists
is whether PINN can be successfully used for a realistic problem in solid mechanics which is dynamic in nature where the
acceleration terms in the governing equations are active. In this work, we use PINN to gauge it’s applicability and effectiveness
for dynamic forward and inverse problems in a linearly elastic setting. For training data, we use the FEM simulation results. Of
particular focus in this study is the use of PINN to create a surrogate model for material identification. This, in particular, is a
major motivation coming from an industrial perspective because there are multiple areas in the industry where solutions to such
problems are computationally expensive and we show that the technology of PINN can circumvent this issue.
The study is organized as follows: in Sec. 2, the methodology of PINN is introduced; Following that, the governing equations
for a linearly elastic problem is discussed in Sec. 3; In Sec. 4, an adaptation of a PINN model to the linearly elastic problem is
shown along with a discussion on the implementation; Next, in Sec. 5, the cantilever problem is introduced and subsequently,
the forward, inverse and the surrogate modeling results are explored in detail with an emphasis on industrial applicability. Also,
in the subsection Sec. 5.4, the extension of the results to a three-dimensional setting is explored. Finally, the conclusions are
discussed in Sec. 6.
∂t u = M u + N (u), (1)
where N is the nonlinear operator and M is the linear operator. Using the neural network graph structure (Kochenderfer and
Wheeler 2019), the linear and nonlinear terms are calculated. For accurate neural network based approximation of the solution,
a loss function has to be minimized (Raissi et al. 2019; Jin et al. 2021). The loss functions are defined as follows.
The equation loss Leqn is given as:
Neqn
1
Leqn = ∑ |∂t u(xen ,tne ) − M u(xen ,tne ) − N (u(xen ,tne ))|2 , (2)
Neqn n=1
where (xe ,t e ) denote the collocation points where equation loss is calculated and Neqn denotes the total number of these colloca-
tion points. The boundary and initial condition loss functions are given as:
1 NBC
LBC = ∑ |G(u(xbn ,tnb ))|2 , (3)
NBC n=1
1 Nini
Lini = ∑ |u(xin , 0) − ue (xin , 0)|2 , (4)
Nini n=1
where (xb ,t b ) are collocation points on the boundaries where the boundary loss is trained and (xi , 0) are collocation points where
the initial condition loss is trained. NBC denotes the number of collocation points on the boundaries and Nini denotes the number
3
of collocation points at initial time. ue (x, 0) is the initial conditions for the set of equations in (1). The function G is a measure
of the deviation of the predicted solution from the imposed boundary conditions (BCs). It takes only positive real values for any
solution field. The final loss that we compute is the loss data Ldata which measures the deviation of the predicted solution from
the exact solution. It is defined as,
Ndata
1
Ldata = ∑ |u(xdn ,tnd ) − ue (xdn ,tnd )|2 , (5)
Ndata n=1
where (xd ,t d ) are collocation points for training the system on a sparse set of interior points and ue denotes the exact solution
which is obtained either from simulation or from experimental data. Ndata is the number of collocation points for training on data
of interior points. In this work, we consider only the data loss and the equation loss. The total loss is thus given by:
Ltot = λ1 Ldata + λ2 Leqn , (6)
where λ1 and λ2 are weight coefficients for the contribution from data and equation losses respectively.
The inputs to the neural network are the spatial and temporal coordinates (x,t). According to Goodfellow et al. (2016), each
network layer output is given as:
zl = σl (zl−1 Wl + bl ) , (7)
where l is the layer number, W represents the weight, b the bias and σl is the activation function. The weights and biases are
optimized so as to minimize the total loss Ltot .
3. Linear elasticity
In this section, we give an overview of the governing equations for linear elasticity. We start with the equation for dynamic
momentum balance
∂ 2 u(x)
∇ · σ(x) + ρb(x) = ρ , x ∈ Ω, (8)
∂t 2
where, σ is the stress tensor field, ρ is the density of the material, b is the body force and u is the displacement. x are the points
inside the domain Ω. The stress tensor field is related to the strain tensor field ϵ(x) by a constitutive relation
1
ϵ(x) = (∇u(x) + ∇uT (x)), x ∈ Ω. (10)
2
We assume plane-strain conditions and no body forces. Therefore, in scalar form the Eq. (8) becomes
∂ σxx ∂ σxy ∂ 2 ux
+ =ρ 2 , (11)
∂x ∂y ∂t
∂ σyy ∂ σxy ∂ 2 uy
+ =ρ 2 . (12)
∂y ∂x ∂t
Likewise, the constitutive relation in Eq. (9) based on the kinematic relation from Eq. (10) becomes
∂ ux ∂ uy
σxx = (λ + 2µ) +λ , (13)
∂x ∂y
∂ uy ∂ ux
σyy = (λ + 2µ) +λ , (14)
∂y ∂x
∂ ux ∂ uy
σxy = µ + . (15)
∂y ∂x
4
where the scaling factors ·c are chosen to be the absolute maximum value of the quantities under consideration for the forward
problem. We assume xc = yc = lc and λc = µc . For forward simulations, we know the value of the material properties. However,
for the inverse problem or surrogate modeling, since we do not have any prior information on the material properties, we have to
assume a value for λc and µc . In a practical scenario, for an inverse problem, for training data, one will perform measurements
on a newly manufactured material and one would know a rough value for the properties which can then be chosen for the scaling.
Also, while performing neural network prediction, the quantities with ·∗ will be scaled back to original range.
where the individual losses based on Eqs. (2) and (5) can be expanded as
∗ ∗
∂ σxx σxy c ∂ σxy uxc ρc lc ∗ ∂ 2 u∗x
Lx = ∗
+ ∗
− ρ (20)
∂x σxxc ∂ y σxxctc2 ∂t ∗ 2 eqn
∗ ∗
σxy c ∂ σxy σyy c ∂ σyy uy c ρc lc ∗ ∂ 2 u∗y
Ly = + − ρ (21)
σxxc ∂ x∗ σxxc ∂ y∗ σxxctc2 ∂t ∗ 2 eqn
σxxc lc ∗ uxc ∗ ∂ u∗ ∂ uy ∗
Lxx = σxx − (λ + 2µ ∗ ) ∗x − λ ∗ ∗ (22)
uy c λc uy c ∂x ∂y eqn
∗
∂ uy uxc ∂ u∗x
σxy c lc ∗
Lxy = σxy − µ ∗ + (24)
u y c λc ∂ x∗ uy c ∂ y∗ eqn
N N
The norm ∥.∥data = N 1 ∑n=1 data 1
|q(xdn ,tnd )|2 where q is a quantity of interest. Likewise, ∥.∥eqn = Neqn eqn
∑n=1 |q(xen ,tne )|2 . Also,
data
α1 to α5 in front of the terms in Eq. 34 are the weight coefficients. As for neural network training, we take two of the useful
5
suggestions provided by (Wang et al. 2023) in our PINN model implementation. Firstly, we use the modified neural network
architecture proposed by (Wang et al. 2021) where the forward pass of a L-layer architecture is defined as
u(x) = hl wl + bl , (29)
For optimization, we use stochastic gradient descent using the Adam optimizer (Kingma and Ba 2017). Other important
parameters for our training are the batch-size, epochs, number of equation points and data points. The batch-size is the number
of training instances in the batch and one epoch is one complete training round covering the entire data-set. For example, if we
have a data-set of size N and a batch-size of Nbatch , then the total number of batches will be N/Nbatch . Therefore, for one epoch,
we do the gradient descent optimization for each batch. Also, in each gradient descent step, the equation points are randomly
chosen from the domain Ω. Hence training through one epoch consists of N/Nbatch number of stochastic gradient descent steps.
Unless specified, for all 2D PINN simulations, we keep the equation loss coefficients (α1 , α2 , α3 , α4 , α5 ) as unity.
The other suggestion from (Wang et al. 2023) we follow is the non-dimensionalization of the governing equations based on the
scalings provided in Eq. (16). This ensures that the output quantities of interest from the neural network are within reasonable
range of values. Our scalings are similar to the work of (Henkes et al. 2022a). Since we use data loss as one of the cost functions
to minimize, we have access to the maximum values of each quantity of interest. As noted in (Henkes et al. 2022a), the scaling
only affects the weighting of the loss terms and not the dynamics of the partial differential equations as long as it is linear. Once
the model is trained, the outputs can be scaled back to their actual values. As for the machine learning software, we prepare codes
from scratch using version 2.0 of the TensorFlow library (Abadi et al. 2016).
Figure 2: Schematic of the two-dimensional (2D) cantilever beam under an oscillating excitation at the free end. For clarity,
blue and orange lines show how the beam moves up and down.
10
Total loss
Total loss
Total loss
1 10 1 10 1
10 2 10 2
10 2
10 3 10 3
10 3
10 4
N 64 H 4 10 4 N 64 H 5 N 64 H 6
10 4
N 128 H 4 N 128 H 5 N 128 H 6
10 5 N 256 H 4 10 5 N 256 H 5 10 5 N 256 H 6
100 101 102 103 100 101 102 103 100 101 102 103
Epoch Epoch Epoch
(a) (b) (c)
Figure 3: Total loss for different number of neurons and layers with 6.76 % training data.
7
102 102
101
101 101
100 100
100
10 1 10 1
10
Total loss
Total loss
Total loss
1
10 2 10 2
10 2
10 3
10 3
10 3
10 4
N 64 H 4 10 4 N 64 H 5 10 4 N 64 H 6
N 128 H 4 N 128 H 5 N 128 H 6
10 5
10 5 N 256 H 4 N 256 H 5 10 5 N 256 H 6
100 101 102 103 100 101 102 103 100 101 102 103
Epoch Epoch Epoch
(a) (b) (c)
Figure 4: Total loss for different number of neurons and layers with 11.52 % training data.
102
101
101 101
100 100
100
10 1 10 1
Total loss
Total loss
Total loss
10 1
10 2 10 2
10 2
10 3 10 3
N 64 H 4 N 64 H 5 10 3
N 64 H 6
10 4 N 128 H 4 10 4 N 128 H 5 N 128 H 6
N 256 H 4 N 256 H 5 10 4
N 256 H 6
10 5 10 5
100 101 102 103 100 101 102 103 100 101 102 103
Epoch Epoch Epoch
(a) (b) (c)
Figure 5: Total loss for different number of neurons and layers with 23.9 % training data.
Table 1: Best set of hyperparameters for different training data % along with the corresponding errors when compared with the
FEM solution.
Num. neurons Num. Layers training data collocation points εux εuy
256 4 6.76 % 500 3.12 × 10−3 5.12 × 10−4
128 4 11.56 % 500 3.07 × 10−3 5.48 × 10−4
128 5 23.9 % 500 2.91 × 10−3 5.2 × 10−4
Figs. 3, 4 and 5. From Fig. 3, for the training data of 6.76 % with batch size of 142, we find the computationally optimal set of
neurons and layers to be (256, 4). Similarly, for 11.52 % training data with batch size of 121, the best set of neurons and layers
are (128, 4). Finally, for 23.9 % training data with batch size of 251, it is (128, 5). Here the % data is shown with respect to
boundary data from the FEM solution. Also, we use 500 random collocation points for the calculation of Leqn at each gradient
descent step. These collocation points are sampled using Latin hypercube sampling (LHS) method (McKay et al. 1979).
10
ux FEM ×10 2 5.9 10
ux PINN ×10 2 5.9 10
|uxFEM uxPINN| ×10 3 3.1
8 8 8
6 1.6 6 1.6 6 2.0
x
x
4
2.8 4
2.8 4
1.0
2 2 2
0
0 25 50 75 100 125 150 175 200
7.2 0
0 25 50 75 100 125 150 175 200
7.2 0
0 25 50 75 100 125 150 175 200
0.0
y y y
10
uy FEM ×10 1 13.7 10
uy PINN ×10 1 13.7 10
|uyFEM uyPINN| ×10 3 15.9
8 8 8
6 9.1 6 9.1 6 10.6
x
x
4
4.6 4
4.6 4
5.3
2 2 2
0
0 25 50 75 100 125 150 175 200
0.0 0
0 25 50 75 100 125 150 175 200
0.0 0
0 25 50 75 100 125 150 175 200
0.0
y y y
10 xx FEM ×10 5 17.3 10 xx PINN ×10 5 17.3 | FEM
xx xx |
PINN ×10 5 4.4
10
8 8 8
6 5.8 6 5.8 6 2.9
x
x
4
5.7 4
5.7 4
1.5
2 2 2
0
0 25 50 75 100 125 150 175 200
17.2 0
0 25 50 75 100 125 150 175 200
17.2 0
0 25 50 75 100 125 150 175 200
0.0
y y y
10 yy FEM ×10 5 3.5 10 yy PINN ×10 5 | FEM
yy yy |
PINN ×10 5 3.8
10
8 8 8
6 1.1 6 1.1 6 2.5
x
x
4
1.3 4 4
1.3
2 2 1.3 2
0
0 25 50 75 100 125 150 175 200
3.6 0
0 25 50 75 100 125 150 175 200
0
0 25 50 75 100 125 150 175 200
0.0
y y y
10 xy FEM ×10 6 14.7 10 xy PINN ×10 6 | FEM
xy xy |
PINN ×10 6 10.8
10
8 8 8
6 5.2 6 5.2 6 7.2
x
x
4
4.2 4
4.2 4
3.6
2 2 2
0
0 25 50 75 100 125 150 175 200
13.7 0
0 25 50 75 100 125 150 175 200
0
0 25 50 75 100 125 150 175 200
0.0
y y y
Figure 6: Contour plots of the quantities of interest at t = 1.72 s for material properties λ = 0.533334 MPa and µ = 0.1 MPa.
Plots in (a) denote FEM solutions. Plots in (b) denote PINN solutions. Plots in (c) denotes the absolute error.
where, nt is the total number of spatial or spatio-temporal points and u represents the quantity of interest. The reason we use
this metric is because the numerical values of the quantities of interest vary by different orders of magnitude. For the best of
hyperparameter sets for their corresponding training data %, Table 1 shows the NRMSE error ε in ux and uy as compared with
the FEM solution. We compute ε at 101 × 11 × 100 spatio-temporal points which corresponds to 101 × 11 sparse spatial grid
points and 100 time instants. We observe that for all three training data sets, the testing space error is of the order 10−3 and
10−4 for the displacements ux and uy which is reasonably small. Errors for training data of 6.76 % is of the same order as that
of other training data sets. So, for the following results in this section we have chosen the training data of 6.76 %. In Fig. 6, we
compare the contour plots of displacements and stresses from the PINN model with the FEM solution at a particular time t = 1.51
s. We observe an excellent agreement between PINN and FEM. Considering we used only a sparse set of training points on the
boundary, it is remarkable that PINN is able to learn the solution at all the other interior spatial locations. Apart from the contour
plots, we are also interested to see the time series of displacements and stresses at particular locations on the boundary of the
beam. We show the same in Fig. 7. We note that the displacements closely match with the FEM solution. As for the stresses even
though it is not a spot on match, the PINN solutions are of the same order of magnitude. Next, we show the spatial NRMSE error
ε in displacements and stresses with respect to time in Fig. 8. We observe that error is low and there is no general trend of error
growth or decay in time. Considering we have shown convincing results from PINN for a sparse training data set of 6.76 %, if
we analogously assume that we have a sparse set of measurement data, we could use this data along with a PINN model to obtain
complete information of the displacements and stresses for all assumed spatio-temporal points for this dynamic system. This
makes practical sense because we usually perform measurements at limited sensor locations due to cost considerations. Another
advantage of the PINN model is that we get solution information at the interior of the structure which has similar accuracy to the
FEM solution.
9
100
10 1
10 2
10 2
0 0
ux
uy
10 2
10 2
10 1
ux FEM uy FEM
ux PINN 100 uy PINN
10 1
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
t t
(a) (b)
10 6 10 5
10 4
10 5
0 0 0
xx
yy
xy
10 5
Figure 7: Line plots of the quantities of interest at (x = 120 mm, y = 8 mm, t = 1.72 s) for material properties λ = 0.533334
MPa and µ = 0.1 MPa. Plots in (a) to (e) show the comparison of displacements and stresses obtained by PINN with respect to
the FEM solutions.
ux
uy
10 2
10 3
xx
10 3
yy
xy
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
t t
(a) (b)
Figure 8: Time series of errors for the material properties λ = 0.533334 MPa and µ = 0.1 MPa. Plot (a) and (b) show the error
in displacements and stresses respectively.
10
100 True
105 105 PINN
10 1
6 × 104
Total Loss
10 2
4 × 104
10 3 3 × 104
104 True
PINN
10 4 2 × 104
100 101 102 103 104 100 101 102 103 104 100 101 102 103 104
Epoch Epoch Epoch
Figure 9: Plot (a) shows the total loss, plots (b) and (c) show the evolution of predicted material properties λ and µ over the
course of training epochs.
Table 3: Error values for different training and testing values of the material parameter µ with respect to FEM solution. Bottom
three rows denote testing space of the parameters.
the geometry is complicated. Secondly, it can result in increased experimental costs because as more data is needed, more sensors
are required. Due to these reasons, data quantity and quality for training the PINN model in this context will be limited. Hence,
we will not expand upon the results of inverse models further in this study. However, the results we obtained in this context could
be exciting for applications based on tomography.
A schematic of the problem is shown in Fig. 12 where the 3D cantilever beam is under oscillating excitation at the free end in
both y and z axes. We consider beam of size 200 mm × 10 mm × 10 mm with x ∈ [0, 200] mm, y ∈ [0, 10] mm , z ∈ [0, 10] mm.
We assume a linearly elastic material with a density of ρ = 0.92 × 10−6 kg/mm3 . The material properties are considered to be
the same as considered for the 2D surrogate model. For the FEM simulation, spatially, the solution is computed on a grid with
24321 nodes of which 8202 nodes belong to the boundary surfaces. The total simulation time is set for 3 seconds.
12
10
ux FEM ×10 2 10.4 10
ux PINN ×10 2 10.4 10
|uxFEM uxPINN| ×10 3 4.3
8 8 8
6 1.8 6 1.8 6 2.9
x
x
4
6.9 4
6.9 4
1.4
2 2 2
0
0 25 50 75 100 125 150 175 200
15.5 0
0 25 50 75 100 125 150 175 200
15.5 0
0 25 50 75 100 125 150 175 200
0.0
y y y
10
uy FEM ×10 1 29.6 10
uy PINN ×10 1 29.6 10
|uyFEM uyPINN| ×10 3 19.0
8 8 8
6 19.7 6 19.7 6 12.7
x
x
4
9.9 4
9.9 4
6.3
2 2 2
0
0 25 50 75 100 125 150 175 200
0.0 0
0 25 50 75 100 125 150 175 200
0
0 25 50 75 100 125 150 175 200
0.0
y y y
10 xx FEM ×10 5 29.8 10 xx PINN ×10 5 23.6 | FEM
xx xx |
PINN ×10 5 17.6
10
8 8 8
6 10.0 6 7.5 6 11.7
x
x
4
9.8 4
8.7 4
5.9
2 2 2
0
0 25 50 75 100 125 150 175 200
29.6 0
0 25 50 75 100 125 150 175 200
24.8 0
0 25 50 75 100 125 150 175 200
0.0
y y y
10 yy FEM ×10 5 18.2 10 yy PINN ×10 5 14.9 | FEM
yy yy |
PINN ×10 5 13.5
10
8 8 8
6 6.1 6 4.6 6 9.0
x
x
4
6.0 4
5.7 4
4.5
2 2 2
0
0 25 50 75 100 125 150 175 200
18.1 0
0 25 50 75 100 125 150 175 200
16.0 0
0 25 50 75 100 125 150 175 200
0.0
y y y
10 xy FEM ×10 6 72.9 10 xy PINN ×10 6 77.0 | FEM
xy xy |
PINN ×10 6 28.9
10
8 8 8
6 34.1 6 42.7 6 19.3
x
x
4
4.6 4
8.4 4
9.6
2 2 2
0
0 25 50 75 100 125 150 175 200
43.3 0
0 25 50 75 100 125 150 175 200
26.0 0
0 25 50 75 100 125 150 175 200
0.0
y y y
Figure 10: Contour plots of the quantities of interest at t = 1.72 s for the material property µ = 0.075 MPa. Plots in (a) denote
FEM solutions. Plots in (b) denote PINN solutions. Plots in (c) denotes the absolute error.
Ldata = ∥u∗x − u∗x d ∥data + ∥u∗y − u∗y d ||data + ∥u∗z − u∗z d ||data + ∥σxx
∗ ∗d
− σxx ∗
∥data + ∥σyy ∗d
− σyy ∥data +
∥σzz∗ − σzz∗ d ∥data + ∥σxy
∗ ∗d
− σxy ∗
∥data + ∥σyz ∗d
− σyz ∗
∥data + ∥σxz ∗d
− σxz ∥data (33)
100
10 1
10 2
10 2
0 0
ux
uy
10 2
10 2
10 1
ux FEM uy FEM
ux PINN 100 uy PINN
10 1
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
t t
(a) (b)
10 5
10 4
10 6
10 5
0 0
xx
yy
xy
10 5
Figure 11: Line plots of the quantities of interest at (x = 120 mm, y = 8 mm, t = 1.72 s) for material property µ = 0.075 MPa.
Plots in (a) to (e) show the comparison of displacements and stresses obtained by PINN with respect to the FEM solutions.
Figure 12: Schematic of the 3D cantilever beam under an oscillating excitation at the free end. For clarity, blue and orange
beams show how the beam moves in 3D at two different instances in time.
∗ ∗ ∗
∂ σxx σxy c ∂ σxy σxz c ∂ σxz uxc ρc lc ∗ ∂ 2 u∗x
Lx = + + − ρ (35)
∂ x∗ σxxc ∂ y∗ σxxc ∂ z∗ σxxctc2 ∂t ∗ 2 eqn
∗ ∗ ∗
σxy c ∂ σxy σyy c ∂ σyy σyz c ∂ σyz uy ρc lc ∂ 2 u∗y
Ly = ∗
+ ∗
+ ∗
− c 2 ρ ∗ ∗2 (36)
σxxc ∂ x σxxc ∂ y σxxc ∂ z σxxctc ∂t eqn
14
∗ ∗
σxz c ∂ σxz σyz c ∂ σyz σzz c ∂ σzz∗ uz c ρc lc ∗ ∂ 2 u∗z
Lz = + + − ρ (37)
σxxc ∂ x∗ σxxc ∂ y∗ σxxc ∂ z∗ σxxctc2 ∂t ∗ 2 eqn
σxxc lc ∗ uxc ∗ ∂ u∗ uy ∂ uy ∗ ∂ uz ∗
Lxx = σxx − (λ + 2µ ∗ ) ∗x − c λ ∗ ∗ − λ ∗ ∗ (38)
uz c λc uz c ∂x uz c ∂y ∂z eqn
∂ u∗y ∂ u∗z
σyz c lc ∗ ∗ uy c
Lyz = σ −µ + (42)
uz c λc yz uz c ∂ z∗ ∂ y∗ eqn
We consider the equation loss coefficients α1 to α9 as unity during the training. We discuss the results from the PINN model
for an unseen or testing material parameters µ. We show in Fig. 13 the comparison of displacements with respect to the FEM
results for µ = 0.075 MPa. We observe a close visual and quantitative match between the PINN and FEM solutions. For ux , the
absolute error is maximum at top right and bottom left sides of the beam. For uy , the absolute error is maximum close to the
mid section of the beam along x direction. For uz it is maximum at the right end of the beam. The reason we choose to show
only displacements instead of the stresses is because of a practical reason that in real industrial experiments, displacements can
be measured accurately using sensors on the material but stress measurements are bit complicated and sometimes not possible.
We compute NRMSE error ε at 51 × 11 × 11 × 100 spatio-temporal points which corresponds to 51 × 11 × 11 sparse spatial
grid points and 100 time instants. In Table 4, we show the numerical value of the errors in the displacements for both training
and testing values of the parameter µ. The bottom three rows comprise of the testing values. We observe the testing errors in
displacements are one order higher as compared to the training ones. This is still acceptable because the testing errors are of the
order O(10−2 ) which is quite low. For an inclusive view of the results, we also show in Fig. 14 the time series of the spatial errors
in displacements and stresses calculated at one random location in the 3D beam. We observe the displacements to be matching
closely. As for the stresses, they are not following the trend of the FEM solution. This is because the neural network has a
bias towards learning lower frequency functions better. However, the predicted stress are of the same order as that of the FEM
solution. For an unseen or testing value of the parameter µ, worthy to note is that the computational cost for such a dynamic 3D
simulation is mere seconds using PINN whereas it takes ∼ 20 minutes using FEM on a standalone Core i7 CPU. Using such a
deep learning framework, if one has to perform N number of simulations for N parameters, that too on a complicated geometry,
one can rely on a trained PINN model to provide extremely fast solutions with reasonable accuracy.
15
10
ux FEM ×10 1 0.7 10
ux PINN ×10 1
10
|uxFEM uxPINN| ×10 3 10.9
8 8 8
6 0.2 6
0.2 6 7.3
y
y
0.4
y
4
0.4 4 4
3.6
2 2 2
0 1.0 0 1.0 0 0.0
0 25 50 75 100 125 150 175 200 0 25 50 75 100 125 150 175 200 0 25 50 75 100 125 150 175 200
x x x
10
uy FEM ×10 1 26.5 10
uy PINN ×10 1 26.5 10
|uyFEM uyPINN| ×10 2 9.2
8 8 8
6 17.6 6 17.6 6 6.2
y
y
4
8.8 4
8.8 4
3.1
2 2 2
0
0 25 50 75 100 125 150 175 200
0.0 0
0 25 50 75 100 125 150 175 200
0
0 25 50 75 100 125 150 175 200
0.0
x x x
10
uz FEM ×10 1 0.0 10
uz PINN ×10 1 0.0 10
|uzFEM uzPINN| ×10 2 8.8
8 8 8
6 4.7 6 4.7 6 5.9
y
y
4
9.4 4
9.4
4
2.9
2 2 2
0
0 25 50 75 100 125 150 175 200
14.1 0
0 25 50 75 100 125 150 175 200
0
0 25 50 75 100 125 150 175 200
0.0
x x x
Figure 13: Contour plots of the quantities of interest in the xy-plane at z = 5 mm and t = 2.29 s for the material property µ = 0.075
MPa. Plots in (a) denote FEM solutions. Plots in (b) denote PINN solutions. Plots in (c) denotes the absolute error.
Table 4: Error values for different training and testing values of the material parameter µ with respect to FEM solution. Bottom
three rows denote testing space of the parameters. (3D problem)
6. Conclusion
In this study, we use the PINN method to solve a dynamic linear elasticity problem. We use state-of-the-art PINN methods
and set up a deep learning framework specifically focused on solving dynamic linear elasticity. We deploy the PINN model for
a forward problem, inverse problem and finally we create a surrogate model. For forward modeling, we perform a systematic
hyperparameter search and we obtain an optimal set for which the total loss is minimal. We show comparison of the PINN results
with respect to the FEM solutions and find that they agree qualitatively and quantitatively. It is to be noted that we use very
minimal data from FEM simulations for training and stress on the analogy that this training data could be from experiments in
an industrial setting. As for the inverse modeling, we show results where we obtain accurate prediction of the parameters λ and
µ. However, keeping in mind the practical reasons in the industry such as minimal sensor data or infeasible measurements at
the interior of the geometry, we don’t delve into a deeper discussion on inverse models. Instead, we would emphasize the use of
PINN in the case of creating a surrogate model.
Initially, we show how PINN could be used as a surrogate model for a 2D dynamic problem where we accurately predict
solutions for an unseen or testing value of the parameter. Later, we create a surrogate model for a 3D dynamic problem. We show
contour and line plots of the results for an unseen parameter in comparison with the FEM solution and we observe an accurate
match in the displacements. The results obtained would significantly reduce the computational cost of searching for the material
parameters iteratively. This is because, once the the PINN model is trained, it takes mere seconds for it to generate solutions
16
100 100
10 1 10 1
10 2
0 10 2 10 2
0 0
ux
uy
uz
10 2 10 2 10 2
10 1 10 1
ux FEM uy FEM uz FEM
ux PINN 100 uy PINN 100 uz PINN
10 1
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
t t t
(a) (b) (c)
10 4 10 5
10 6
10 5
10 6
0
xx
yy
0
zz
0
10 5
10 6
xx FEM yy FEM 10 6 zz FEM
10 4
xx PINN yy PINN zz PINN
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
t t t
(d) (e) (f)
10 5
10 5 10 6
10 6
0
xy
xz
yz
0 0
10 6
Figure 14: Line plots of the quantities of interest at (x = 120 mm, y = 8 mm, z = 5 mm, t = 2.29 s) for material property
µ = 0.075 MPa. Plots in (a) to (e) show the comparison of displacements and stresses obtained by PINN with respect to the FEM
solutions.
for an unseen parameter. The deep learning framework introduced in this study to solve such dynamic problems could also be
extended to solve dynamic problems beyond the linearly elastic regimes.
For future work, we would be applying this PINN framework on problems of industrial relevance, most of which involve
complex geometries and random excitation. Another possible challenge is to add material nonlinearity to the present problem
and still use such a framework to create surrogate models. In summary, in this work, we have shown the successful application of
a deep learning framework based on PINN to a realistic dynamic problem in solid mechanics. In doing so, we show how PINN
can circumvent high computational costs associated with numerical simulations in such problems.
Acknowledgements
The authors thank the computing resources and support provided by the CI/PDP Deep Learning Platform Team at Robert Bosch
GmbH, Renningen, Germany. The authors acknowledge the help of Ponsuganth Ilangovan P for his inputs in setting up the FEM
simulations.
17
REFERENCES
M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, M. Kudlur, J. Levenberg, R. Monga,
S. Moore, D. G. Murray, B. Steiner, P. Tucker, V. Vasudevan, P. Warden, M. Wicke, Y. Yu, and X. Zheng. TensorFlow: A system for
Large-Scale machine learning. In 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), pages 265–283,
Savannah, GA, Nov. 2016. USENIX Association. ISBN 978-1-931971-33-1.
H. Adeli and C. Yeh. Perceptron learning in engineering design. Computer-Aided Civil and Infrastructure Engineering, 4(4):247–256, 1989.
D. Anton and H. Wessels. Physics-informed neural networks for material model calibration from full-field displacement data, 2023.
S. L. Brunton, B. R. Noack, and P. Koumoutsakos. Machine learning for fluid mechanics. Annual Review of Fluid Mechanics, 52(1):477–508,
2020.
S. Cai, Z. Wang, F. Fuest, Y. J. Jeon, C. Gray, and G. E. Karniadakis. Flow over an espresso cup: inferring 3-d velocity and pressure fields
from tomographic background oriented schlieren via physics-informed neural networks. Journal of Fluid Mechanics, 915:A102, 2021.
F. S. Costabal, S. Pezzuto, and P. Perdikaris. δ -pinns: physics-informed neural networks on complex geometries, 2022.
G. Elsayed, S. Shankar, B. Cheung, N. Papernot, A. Kurakin, I. Goodfellow, and J. Sohl-Dickstein. Adversarial examples that fool both
computer vision and time-limited humans. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors,
Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018.
A. Esteva, R. A., and R. e. a. B. A guide to deep learning in healthcare. Nature Medicine, 25:34–29, 2018.
H. Fujiyoshi, T. Hirakawa, and T. Yamashita. Deep learning-based image recognition for autonomous driving. IATSS Research, 43(4):
244–252, 2019. ISSN 0386-1112.
J. Ghaboussi, J. H. Garrett, and X. Wu. Knowledge-based modeling of material behavior with neural networks. Journal of Engineering
Mechanics, 117(1):132–153, 1991.
D. González, F. Chinesta, and E. Cueto. Learning corrections for hyperelastic models from data. Frontiers in Materials, 6, 2019. ISSN
2296-8016.
I. J. Goodfellow, Y. Bengio, and A. Courville. Deep Learning. MIT Press, Cambridge, MA, USA, 2016. http://www.deeplearningbook.org.
S. Grigorescu, B. Trasnea, T. Cocias, and G. Macesanu. A survey of deep learning techniques for autonomous driving. Journal of Field
Robotics, 37:362–386, 2019.
E. Haghighat, M. Raissi, A. Moure, H. Gomez, and R. Juanes. A physics-informed deep learning framework for inversion and surrogate
modeling in solid mechanics. Computer Methods in Applied Mechanics and Engineering, 379:113741, 2021. ISSN 0045-7825.
E. Haghighat, S. Abouali, and R. Vaziri. Constitutive model characterization and discovery using physics-informed deep learning.
Engineering Applications of Artificial Intelligence, 120:105828, 2023. ISSN 0952-1976.
A. Henkes, H. Wessels, and R. Mahnken. Physics informed neural networks for continuum micromechanics. Computer Methods in Applied
Mechanics and Engineering, 393:114790, 2022a. ISSN 0045-7825.
A. Henkes, H. Wessels, and R. Mahnken. Physics informed neural networks for continuum micromechanics. Comput. Methods Appl. Mech.
Engrg., 393:114790, 2022b.
Z. Hu, A. D. Jagtap, G. E. Karniadakis, and K. Kawaguchi. Augmented physics-informed neural networks (apinns): A gating network-based
soft domain decomposition methodology. Engineering Applications of Artificial Intelligence, 126:107183, 2023. ISSN 0952-1976.
A. D. Jagtap and G. E. Karniadakis. Extended physics-informed neural networks (xpinns): A generalized space-time domain decomposition
based deep learning framework for nonlinear partial differential equations. Communications in Computational Physics, 28(5):2002–2041,
2020. ISSN 1991-7120.
A. D. Jagtap, Z. Mao, N. Adams, and G. E. Karniadakis. Physics-informed neural networks for inverse problems in supersonic flows. Journal
of Computational Physics, 466:111402, 2022. ISSN 0021-9991.
X. Jin, S. Cai, H. Li, and G. E. Karniadakis. Nsfnets (navier-stokes flow nets): Physics-informed neural networks for the incompressible
navier-stokes equations. Journal of Computational Physics, 426:109951, 2021. ISSN 0021-9991.
V. Kag, K. Seshasayanan, and V. Gopinath. Physics-informed data based neural networks for two-dimensional turbulence. Physics of Fluids,
34(5):055130, 05 2022. ISSN 1070-6631.
18