Professional Documents
Culture Documents
Introduction To Continuum Mechanics Lecture Notes: Jagan M. Padbidri
Introduction To Continuum Mechanics Lecture Notes: Jagan M. Padbidri
Lecture Notes
Jagan M. Padbidri
Fall 2017
Chapter 1
Essential mathematics
Since our motivation is to describe physical phenomena using vectors, we will confine our-
selves to real valued vectors. The same holds true for scalars as well, i.e., we are interested in
real valued scalars. The set of all possible such scalars thus forms the real value space R. We
also define the set of all real valued vectors to form the space V. Both these spaces are linear
spaces and would need to satisfy certain conditions. Before listing the conditions, we will follow
that vectors will be designated using a bold face font or an underscore and unit vectors will be
identified with a carat (ˆ) symbol. Real valued vectors belonging to the linear vector space V
will need to satisfy the following properties.
Additive commutativity:
For any arbitrarily chosen vectors u, v ∈ V,
u+v=v+u
Here, it is also implied that the resultant sum is a vector and is also a member of the vector
space V. Thus, the addition of vectors is an operation that takes two elements of the vectors
space V and maps the resultant, also, on to the space V, i.e., this operation can be described
as V × V → V. The ‘×’ represents a generic operation, not mathematical multiplication.
Additive associativity:
For any arbitrarily chosen vectors u, v, w ∈ V,
(u + v) + w = u + (v + w)
This is also a V × V → V operation.
1
Zero element:
There exists an element 0 ∈ V such that
u+0=u
Negative element:
For every u ∈ V, there exists a v ∈ V such that
u + v = 0 =⇒ v = -u
u · u = |u||u| = |u|2
√
=⇒ |u| = u · u (1.1)
The dot product is an operation on two vectors that results in a scalar, i.e., it is a V × V → R
operation. Extending the possible values assumed by θ,
u · v = 0 =⇒ u ⊥ v (θ = π/2)
2
We note the dot product also satisfies the following conditions
u·v=v·u Commutativity
u · (v + w) = u · v + u · w Distributivity
α(u · v) = u · (αv) Associativity with scalars
Figure 1.1: (a)Dot product of two vectors and (b) Cross product of two vectors.
u × v = −v × u Anti-Commutativity
u × (v + w) = u × v + u × w Distributivity
α(u × v) = u × (αv) Associativity with scalars
Formation of a basis
The general rules for addition of vectors belonging to linear vector space can be extended to
3
an arbitrary set of vectors, i.e., the sum of an arbitrary set of vectors that belong to V should
result in a vector belonging to V. This can be written in a general form as follows. Given a set
of scalars (α1 , α2 , . . . , αn ) ∈ R and a set of vectors (u1 , u2 , . . . , un ) ∈ V,
α1 u1 + α2 u2 + · · · + αn un = v ∈ V
If for an arbitrary set of vectors (u1 , u2 , . . . , un ) ∈ V, we can find a set of scalars (α1 , α2 , . . . , αn ) ∈
R, not all zero such that the above relation is satisfied, the set (u1 , u2 , . . . , un ) is said to be lin-
early dependent.
Now, if we have a set of linearly independent vectors, (u1 , u2 , . . . , un ) and we chose a set of
scalars, (α1 , α2 , . . . , αn ), not all 0, then
α1 u1 + α2 u2 + · · · + αn un 6= 0
Rather,
α1 u1 + α2 u2 + · · · + αn un = v ∈ V
There are two important conclusions that can be drawn from this. First, since (α1 , α2 , . . . , αn )
are chosen arbitrarily and v ∈ V, this means that for any v ∈ V, a set of (α1 , α2 , . . . , αn ) can be
obtained to satisfy the above relation. Second, for a given v, the set (α1 , α2 , . . . , αn ) is unique
since (u1 , u2 , . . . , un ) is linearly independent. This means that any arbitrary v ∈ V can be
desribed using the linearly independent set (u1 , u2 , . . . , un ), i.e., (u1 , u2 , . . . , un ) acts as a basis
for the description of elements in V. Such a set of vectors are called the basis vectors for the
space V.
The most commonly used set of basis vectors are the ones used to describe vectors in
the cartesean coordinate system. While a vector space is not limited by dimension, since the
cartesean system describes 3D Euclidean space (R3 ), there are three basis vectors in this system.
We commonly refer to them as the x, y and z axes. Since these are basis vectors, any vector in
(R3 ) can be represented as a linear combination of these three basis vectors.
To simplify determining the set of scalars required to express an arbitrary vector in terms of
the basis vectors, we usually impose the condition that the basis vectors are of unit magnitude,
i.e., the vectors along the x, y and z directions are in fact x̂, ŷ and ẑ or î, ĵ and k̂. Also, the
vectors that form the basis are orthogonal to each other, i.e., î · ĵ = ĵ · k̂ = î · k̂ = 0. Hence,
this basis is called an Orthonormal basis. Now, any vector v ∈ R3 space can be expressed as a
4
Figure 1.3: Example of a basis in R3 .
v = vx î + vy ĵ + vz k̂ (1.2)
where vx , vy and vz are the components of v along î, ĵ and k̂ and represent the ‘weight’ of the
vector along said directions. It follows that
Note that it is not necessary that a vector space has a unique basis. It can have infinitely
many basis. However, the representation of any vector v ∈ R3 is unique for that specific basis.
We now revisit the vector operations previously addressed. Consider arbitrary vectors u and v
∈ R3 . The scalar product of the vectors can be written as
5
Index notation
The above examples illustrate that even simple vector operations are rather cumbersome in rep-
resentation. We now introduce the notion of using indices to simplify representation of vectors
and avoid writing lengthy expressions. Since each vector is represented using the basis vectors,
we need to first address representing the basis vectors themselves in the indicial notation. We
replace the set of basis vectors (x̂, ŷ, ẑ) or (î, ĵ, k̂) with the set (ê1 , ê2 , ê3 ). This set can now be
generically represented as êi where i = 1, 2, 3. Note that the index i serves as a placeholder to
assume values 1, 2 or 3 and has no relation to î. We can express a vector as
v = vx î + vy ĵ + vz k̂
= v1 ê1 + v2 ê2 + v3 ê3
3
X
= vi êi
i=1
Note that the choice of using i as the index is arbitrary. Any other alphabet could be used as
an index, per convenience. We have already specified that the index i is not related to î. As
we have detailed previously, êi actually represents a set of three vectors (ê1 , ê2 , ê3 ) or (î, ĵ, k̂).
Similarly, êj represents the set of three vectors, albiet the exact number that i and j assume
can be arbitrary.
Now, consider the simple operation of finding the dot product of two unit vectors êi and
êj . We have already addressed the nature of the dot product when considering vectors from
an orthonormal triad. Just to summarize, êi and êj , each represent the relevant unit vector
depending on the value of the index and we are interested in finding êi · êj . Since êi and êj each
represent three vectors, we have a total of nine possibilities. They are
6
i = 3, j = 2; êi · êj = ê3 · ê2 = 0
i = 3, j = 3; êi · êj = ê3 · ê3 = 1
The use of indicial notation allows us to account for all of these possibilities into the compact
form êi · êj , of course with the knowledge that i and j could take values between 1 and 3. We
can also make another, rather obvious observation from the above set of equations that
êi · êj = 1 if i = j
=0 otherwise
We simplify this by introducing a more convenient symbol called the Kronecker delta given
by
Note that δij has two indices i and j, each of which could take three possible values, i.e., δij
can account for all the nine possibilities shown above. Of course, of these possibilities, only
three have non-zero values (i = j = 1; i = j = 2; i = j = 3). The summation convention for the
repeated indices that we have previously introduced, holds good here as well i.e.,
δii = δ11 + δ22 + δ33 = 3
We note that the notion of finding the dot product of basis vectors can be used directly to
find the dot product of any two vectors. Since any arbitrary vector can be expressed in terms
of a set of scalar components and the basis vectors, we can use the properties of linear vector
spaces as,
u · v = ui êi · vj êj
= ui vj êi · êj
= ui vj δij
The above expression, ui vj δij implies that it is a summation over the repeated indices i and j.
Since both i and j can take values of 1, 2 or 3, the terms of the summation can be written as
7
ui vj δij = u1 v1 + u2 v2 + u3 v3
=⇒ ui vj δij = ui vi
Comparing the respective terms, we get vj δij = vi
Alternatively, let us consider just the term vj δij . For any given value of the index i, vj δij
represents summation over three possible values of j, i.e., vj δij is the sum of three terms with
j taking on values 1, 2 and 3. In these three terms, there will be only one instance when
both i and j have the same value. For the remaining two terms, i 6= j =⇒ δij = 0 and
consequently, those two terms vanish! The only non-zero term in the summation corresponds
to when i = j =⇒ δij = 1 for that particular term. Thus, the summation vj δij condenses to
just one term, and i = j for that term and we can write
vj δij = vi (1.8)
Thus, δij acts on specific components of v and effectively ‘flips’ the index of that component by
rendering the remaining terms of the summation to 0. The dot product of two vectors can be
written as
u · v = ui vj δij
=⇒ u · v = ui vi (1.9)
Since the above expression has a repeated index, we can write out the summation over that
index and verify that the indicial notation does represent the dot product
u · v = ui vi = u1 v1 + u2 v2 + u3 v3
ui and vj are components of the vectors u and v i.e., they are scalars. The challenge lies in
describing the vector product êi × êj . This is seemingly more complex than describing the dot
product of two basis vectors since the result of a dot product is a scalar quantity, but we know
that the result of a cross product is a vector, subject to the right hand rule. Let us write down
the possibilities for êi × êj .
i = 1, j = 1; ê1 × ê1 = 0
i = 1, j = 2; ê1 × ê2 = ê3
i = 1, j = 3; ê1 × ê3 = −ê2
8
i = 3, j = 2; ê3 × ê2 = −ê1
i = 3, j = 3; ê3 × ê3 = 0
We observe certain patterns in the result of the vector product of two basis vectors. When
i = j, the vector product is 0. For the instances when i 6= j and the cross product is non-
zero, the indices i and j yield a vector which corresponds to the third index, k 6= i, j, i.e.,
ê2 × ê3 = ê1 , for example. Also, specific sequences of i and j produce a positive vector in
the right handed system and certain other sequences produce a negative vector. For example,
when i and j belong to the sequence 1, 2, 3, i.e., (i, j = 1, 2; i, j = 2, 3; i, j = 3, 1), the result
is a positive vector along êk , and when i and j are in the cyclically opposite sequence (3, 2,
1), i.e., (i, j = 2, 1; i, j = 3, 2; i, j = 1, 3), the result is a negative vector along êk . This can be
summarized as
êi × êj = ijk êk (1.10)
where ijk is called the permutation symbol whose value is given by
w = u × v = ui êi × vj êj
= ui vj êi × êj
= ui vj ijk êk
=⇒ w = u × v = ijk ui vj êk (1.11)
We have mentioned in the previous section that vectors have both magnitude and direction
and have seen the representation of vectors in an orthonormal triad of basis vectors. However,
we have not specifically addressed the uniqueness of this basis. It is rather obvious that the
three dimensional real Euclidean space (R3 ) we live in does not have a unique set of basis vectors
i.e., one can define a set of basis vectors per convenience and can also ‘switch’ between the bases
so defined. Note that we are referring to changing the basis for a given vector, not changing or
altering the vector itself. This is reflected as a change in the representation of the vector, but
not the vector itself. For example, if we have a vector v and two different bases êi and ēˆj , the
above argument can be summarized as
This still gives no information on relating vi and v̄i or even êi and ēˆi . We will address this by
formulating the transformation of a vector from one basis to another.
9
Consider a basis êi = (ê1 , ê2 , ê3 ) that has been rotated arbitrarily to obtain a new basis
ēˆi = (ēˆ1 , ēˆ2 , ēˆ3 ), as shown in Fig. 1.5. Note that both the bases are right handed orthonormal
triads. Since the basis vectors are not perfectly alligned with each other, each of the ‘old’
basis vectors, êi will have a non-zero component along each of the ‘new’ basis vectors, ēˆi . The
magnitudes of these components can be obtained from the dot products of the relevant basis
vectors, i.e., we can express each of the ‘new’ basis vectors as if they were an arbitrary vector in
the ‘old’ basis and vice versa, i.e., each of the ‘old’ basis vectors can be expressed as an arbitrary
vector in the ‘new’ basis.
ēˆ1 = (ēˆ1 · ê1 )ê1 + (ēˆ1 · ê2 )ê2 + (ēˆ1 · ê3 )ê3
= cos(ēˆ1 , ê1 )ê1 + cos(ēˆ1 , ê2 )ê2 + cos(ēˆ1 , ê3 )ê3
= cos θ11 ê1 + cos θ12 ê2 + cos θ13 ê3
=⇒ ēˆ1 = a11 ê1 + a12 ê2 + a13 ê3
It is emphasized here that one needs to be attentive to the conventions being used to describe
the angle between the ‘new’ and the ‘old’ basis. Here θ represents the angle between one ‘new’
basis vectors and one ‘old’ basis vector. The first index for θ represents the ‘new’ basis, ēˆi and
the second index represents the ‘old’ basis, êi , i.e., θ23 represents the angle between ēˆ2 and ê3 .
The other basis vectors in the ēˆi basis follow a similar mathematical pattern with respect to the
vectors belonging to the êi basis. Thus, the relations between the two bases can be expressed
as
ēˆ1 a11 a12 a13 ê1
ˆ
ē2 = a21 a22 a23 ê2
ê1 = cos(ēˆ1 , ê1 )ēˆ1 + cos(ēˆ2 , ê1 )ēˆ2 + cos(ēˆ3 , ê1 )ēˆ3
= cos θ11 ēˆ1 + cos θ21 ēˆ2 + cos θ31 ēˆ3
ê1 = a11 ēˆ1 + a21 ēˆ2 + a31 ēˆ3
=⇒ êi = aji ēˆj
10
Now, we know that the ‘new’ basis vectors, ēˆi form an othonormal triad, i.e., ēˆi · ēˆj = δij . We
can write
This yields us the very important observation that the transpose of the transformation matrix
is its inverse, i.e., the transformation matrix is orthogonal.
Now, consider a vector that is represented in both the bases - v̄ in the ēˆi basis and is
¯
equivalent to v in the êi basis. Using the previously obtained property of the transformation
matrix,
v̄ = v̄j ēˆj ≡ v = vi êi = vi aji ēˆj
¯
By comparing the relevant coefficients,
The idea that a vector has not changed or varied, i.e., remained invariant, implies that
the representation and method of quantification of a vector have not changed. For example,
consider the representation of the vector described above in both the bases, êi and ēˆi . The
√ √
magnitude of the vector in the basis êi is given by |v| = v · v = vi · vi and in the basis ēˆi
√ √
is |v̄| = v̄ · v̄ = v̄i · v̄i . It is straighforward to prove that the transformation of components
¯ ¯ ¯
between the two vectors, as derived above ensures that the magnitude is the same. However,
more fundamental is the expression for calculating the magnitude of the vector. We see that the
method/ expression to calculate the magnitude remains the same irrespective of the basis being
used i.e., the expression is invariant to the basis. This is the fundamental argument towards
the invariance of a vector with respect to the basis. The reader is advised to repeat the process
to explore the other quintessential property of a vector, its direction. Of course, since the basis
11
has been changed, so will the orientation of the vector described in that basis. Thus, the in-
variance of the direction of a vector needs to be verified using another arbitrarily defined vector.
We have discussed in detail the transformation of vectors from one basis to another. This
can be easily extended to describe rotation of vectors in the same basis. Instead of keeping our
vector fixed and rotating the basis vectors to a new position, we can keep the basis fixed and
define a ‘new’ vector by rotating the existing vector about the basis vectors. Determining the
components of this ‘new’ vector in the same basis is analogous to the previous case where the
vector was fixed and the basis was rotated. Thus, the transformation matrix defined above also
doubles up as a rotation matrix. Such an operation maps a vector from an ‘initial’ position in
the basis to a ‘final’ position/ state. Thus, the rotation matrix acts as a linear operator that
maps one vector into another.
Introduction to Tensors
The idea of mapping one vector to another using linear operators is common place, for
eg., the scalar mass is used to map the force and acceleration vectors. However, the spectrum
of mapping one vector to another has been left incomplete i.e., we have considered the more
convenient examples so far. Mapping the force and acceleration vectors takes place through a
physically measurable scalar and is very relavant physically, with both vectors oriented along
the same direction. We have discussed mapping one vector to another using the rotation matrix
with the directions of the two vectors being different, but the rotation matrix itself does not
correspond to any physically measurable quantity. We now introduce quantities which map/
transform vectors and have physical significance - Tensors.
Consider the following description of the angular momentum of a particle. Let the particle
have a mass m, position r = r1 ê1 + r2 ê2 and angular velocity ω = ω3 ê3 . We are familiar that
the angular momentum vector of the particle is given by h = Iω = mr×v, where v = ω×r is
the velocity of the particle. With a basic check, we can see that for r in the (ê1 , ê2 ) plane and
ω along ê3 , the velocity, v also lines in the (ê1 , ê2 ) plane. The angular momentum vector, h
= mr×v is oriented along ê3 . Thus, the role of the moment of inertia, I is analogous to mass
m in F = ma i.e., it maps ω to h with both vectors along the same direction. We note that
there exist significant differences between both cases since the force and acceleration vectors are
defined with respect to the origin of the basis, while angular momentum and angular velocity
can be defined about an arbitrary point or an axis that does not align with any individual basis
vector. Nevertheless, angular momentum and angular velocity are vectors which are mapped
onto each other by the moment of inertia.
Now, consider that the particle has the same angular velocity, ω = ω3 ê3 , but its position
vector is given by r = r1 ê1 + r2 ê2 + r3 ê3 . We can deduce that v = ω×r is still parallel to the
(ê1 , ê2 ) plane. However, the angular momentum vector, h = mr×v will not be along ê3 since r
does not lie in the (ê1 , ê2 ) plane. So here, the moment of inertia maps ω to h which are along
different directions. This deviates from our previous observation of I being analogous to m and
12
Figure 1.6: Schematic of a rotating mass and associated vectors.
one can deduce that I is not a mere scalar. Let us analyse the expression for angular momentum
in detail. We know,
h = mr × v
= mr × (ω × r)
= m(r · r)ω − m(r · ω)r
Similar to the rotation of a vector, the components of the angular momentum and angular
velocity vectors can be expressed as column vectors and the inertia in a matrix form as
h1 m(r2 2 + r3 2 ) −m(r1 r2 ) −m(r1 r3 ) ω1 I11 I12 I13 ω1
2 2
h2 = −m(r2 r1 ) m(r1 + r3 ) −m(r2 r3 ) ω2 = I21 I22 I23 ω2
At first glance, the structure of the above equation looks very similar to that of transforma-
tion of vectors between bases. However, the inertia ‘matrix’, I and the transformation matrix,
A represent two fundamentally different quantities. The transformation matrix operates on a
vector to result its components in a new basis. Note that the composition of the transformation
matrix is independent of the vector it operates on i.e., the components aij are not derived from
the components of the vector they act on, but are derived from the two bases involved. The
inertia ‘matrix’, however, is explicitly derived from the position of the particle (r) and its mass.
Consequently, any change in r is summarily reflected in I. Thus, I has a more sound physical
basis and significance than A and represents a property/ state of the particle in the current
13
basis, in addition to performing the mathematical function of mapping ω to h. Such quanities
are called as Tensors. However, this does not imply that A is not a tensor. Indeed it is one
and belongs to a class of orthogonal tensors.
We have presented arguments that the components of the Inertia Tensor are derived from
components of the position vector, r. This implies that if one changes the basis such that the
components of r are altered, then the components of I should also change, i.e., the inertia tensor
(and tensors in general) are defined in a specific basis, similar to vectors.
Further we note that each component of I is associated with two basis vectors. For example,
consider how the second row of the I matrix operates. We have h2 = I21 ω1 + I22 ω2 + I23 ω3 . We
know that h2 represents the component of the angular momentum in the ê2 direction and ω3
represents the component of the angular velocity in the ê3 direction. The component I23 is thus
mapping a component along ê3 to a quantity along ê2 , i.e., each component of I is associated
with two basis vectors. It may be argued that I22 deals exclusively with the ê2 direction. The
relevant interpretation here is that I22 maps one component of the vector ω to one component
of the vector h - they just happen to be along the same direction! Thus, each component of the
tensor I is associated with two basis vectors and consequently, would require two basis vectors
to describe/ represent it.
Also evident is that a vector, which is associated with one basis vector for each component,
can be represented by a column, whereas I requires a two dimensional structure (matrix) for its
representation since each component of I is associated with two basis vectors. Thus, we refer
to I as a second order tensor. Similarly, vectors are referred to as first order tensors and scalars
are zero order tensors. Since we have identified inertia to be a second order tensor, we represent
it as I.
We have previously seen operations such as the dot product or the scalar product in which
a vector is reduced to a scalar, i.e., a first order tensor is reduced to a zero order tensor. Here,
we introduce the notion of tensor product or dyadic product represented by the ⊗ symbol. In
this operation, two first order tensors (vectors) are operated on to form a second order tensor
given by êi ⊗ êj = êi êj . The dyadic or tensor product of two vectors can be written as
where Tij are the components of the second order tensor T formed by the tensor product of u
and v. The second order tensor has two free indices represented by the basis vectors êi and êj
and its components are represented as a matrix. In the present case, the the inertia tensor can
now be written as
where I is the identity tensor given by I = ê1 ⊗ ê1 + ê2 ⊗ ê2 + ê3 ⊗ ê3 . Note that all tensors need
not be defined based on position vectors. There are many instances where second and higher or-
14
der tensors are used to describe a specific property of a material, independent of position vectors.
Once the Inertia tensor has been formed, it acts upon the angular velocity vector to result
the angular momentum vector. We have addressed the fact that the Inetia tensor is a second
order tensor (I) and the angular velocity vector is a first order tensor (ω). This implies that I
operates on ω to reduce the number of free indices i.e., the operation beween I and ω is a scalar
or dot product. This can be written as
h=I·ω
= Iij êi êj · ωk êk
= Iij ωk êi êj · êk
= Iij ωk êi δjk
=⇒ hi êi = Iij ωj êi
Properties of Tensors
We note from the definition of the tensor I that when represented in a matrix, I is symmetric,
i.e., Iij = Iji . If, for some tensor A, Aij = −Aji then that tensor is said to be skew-symmetric.
To satisfy the requirements of skew-symmetry, the diagonal elements of such a tensor are 0.
15
In general, it is not necessary for a tensor to satisfy the requirements of symmetry or skew-
symmetry. However, any tensor can be represented as a sum of a symmetric and a skew-
symmetric tensor, i.e., any tensor C can be written as
C=A+B
!
C + CT
Aij = Aji is a symmetric tensor, where A =
2
!
C − CT
Bij = −Bji is a skew-symmetric tensor, where B =
2
We have seen that the Inertia tensor maps the angular velocity vector into the angular mo-
mentum vector. We summarize that the angular momentum is defined about a point that is
considered to be the origin and the angular velocity is a vector that passes through this origin.
The position vectors are also defined with respect to the origin. Also, the Inertia tensor maps
the angular velocity vector to the angular momentum vector such that they need not be along
the same direction. The formula for the angular momentum is given by
The angular momentum vector, h is comprised of two different parts, one along ω and the
other along r. It is obvious that when ω ⊥ r, the part of h along r goes to 0 and hkω. We
can present these arguments about the angle between ω and r and the associated orientation of
h for the idealized example we have considered, i.e., a point mass which is described by a single r.
For a finite dimensional body, i.e., something which cannot be approximated to a point mass
and thus cannot be described by a unique position vector r, the above generalizations would not
hold. The Inertia tensor for such a body would have to be obtained by considering the moments
of inertia of infinitismal elements of the body and integrating them. For such an object, the
only mathematical possibility for h to lie along ω is for I to be a purely diagonal matrix, i.e.,
the off-diagonal terms should vanish. For each infinitismal elemet considered, the components
of the Inertia tensor of that element are functions of the components of the position vector of
that infinitismal element, i.e., I12 is a function of r1 and r2 , I13 is a function of r1 and r3 etc.
We recall that the components of any vector depend on the basis in which they are described.
Thus, one may select a basis to represent the finite dimensional body such that the off-diagonal
components of the Inertia tensor for a given infinitismal element may be non-zero, but when
that component of the inertia tensor is integrated over the domain of the body, it reduces to
zero. In such a case, the inertia tensor for the entire body would be represented by a diagonal
matrix and consequently, h k ω. Mathematically, this condition can be written as
h = I · ω = λω
I · ω − λω = 0
I · ω − λI · ω = 0
(I − λI) · ω = 0 (1.13)
We know that the solution for this exists when
16
I − λI = 0 (1.14)
The set of vectors which from the basis in which this condition is satisfied are called as Eigen-
vectors and the values of λ thus obtained are called as Eigenvalues. We can solve for these
by expressing the above equation in a matrix form as
I − λI = 0
=⇒ Iij − λIij = 0
I11 − λ 0 0
0 I22 − λ 0 =0
0 0 I33 − λ
which results in a cubic equation in λ i.e., three possible Eigenvalues, λI , λII and λIII . For each
of these Eigenvalues, there exixts a vector which satisfies Eqn. (1.13) and is the corresponding
Eigenvector, i.e., for λ = λI , there exists a vector ω I such that
(I − λI I) · ω I = 0
The set of vectors for all the three possible values of λ are called as the Eigenvectors. Physically,
this means that around this set of vectors/ axes the mass distribution is symmetric. For a tensor
whose components are real valued and symmetric, the Eigenvectors form an orthogonal triad.
Consider a particle that moves with progression of time i.e., the position vector of the particle
changes with time and can thus be represented as a function of time as r = r(t). Let S be the
path that the particle traces. We consider the particle at two instances of time, t and t + ∆t.
Then the velocity of the particle is given by the derivate of the particle position as
∆r
v = lim
∆t→0 ∆t
17
Figure 1.7: Differentiation of the position vector.
i.e., the magnitude of the velocity vector is the rate at which distance is covered, but it is
oriented tangential to the path along which the particle is travelling. If we express êt in terms
of the basis vectors êi , then we can express the velocity as v = vi êi . Note that since êi s are
independent of time, t, this implies that the velocity can be integrated to obtain a general
expression for the position of the particle as r = ri (t)êi .
Gradient operator
φ = φ(x, y, z) = φ(x1 , x2 , x3 )
Note that the position vector of point P is given by P = xê1 + yê2 + zê3 ≡ x1 ê1 + x2 ê2 + x3 ê3 .
φ has been defined to be a continuous function dependent on all the three coordinates. Thus,
the total differential for φ is given by
∂φ ∂φ ∂φ
∆φ = δx + δy + δz
∂x ∂y ∂z
∂φ ∂φ ∂φ
= δx1 + δx2 + δx3
∂x1 ∂x2 ∂x3
∂φ
= δxi
∂xi
We note that the δxi s are essentially components of an infinitismal vector, δn, i.e., δn = δx1 ê1 +
δx2 ê2 + δx3 ê3 . We also note that the terms involving the partial derivatives could possibly
have directional dependence. We know that φ is a scalar valued function, i.e., it does not have
∂φ
components in specific directions. However, to evaluate , one finds the rate of change of
∂x1
φ only as x1 changes, which would obviously be along ê1 . Thus, the set of partial derivatives
18
evaluated are actually direction dependent and represent the rate of change of φ along the
directions of the basis vectors. The above equation for change of φ can then be written as
∂φ ∂φ ∂φ
∆φ = δx1 + δx2 + δx3
∂x1 ∂x2 ∂x3
∂φ ∂φ ∂φ
= ê1 + ê2 + ê3 · (δx1 ê1 + δx2 ê2 + δx3 ê3 )
∂x1 ∂x2 ∂x3
The function φ is a scalar function and does not have any components, implying that the
directional dependence in the above relation is derived from the operator determining the partial
derivative of φ. This operator is represented using the symbol ∇ and referred to as the Del or
¯
the grad or the nabla operator. Mathematically, it represents the Gradient of φ. Thus,
∂ ∂ ∂ ∂φ ∂φ ∂φ ∂φ
∇= ê1 + ê2 + ê3 such that ∇φ = ê1 + ê2 + ê3 = êi (1.15)
¯ ∂x1 ∂x2 ∂x3 ¯ ∂x1 ∂x2 ∂x3 ∂xi
The gradient of φ is rate at which the scalar function changes in space. The expression for
∆φ indicates that the total change in φ is equal to the rate of change of φ dotted with the
direction of interest. More specifically, the above equation would represent the change of φ
along a specific direction n given by
∆φn = ∇φ · δn (1.16)
¯
where δn has not been specified to be a unit vector. From the definition of the dot product, the
value of the the above dot product is maximum when δn k ∇φ. This means that ∆φ is maximum
¯
along the direction of ∇φ, i.e., the gradient of a field is a vector that gives the magnitude and
¯
direction of the greatest change for that field. The vector δn can be written in terms of a unit
vector as δn = ∆S n̂ where ∆S is the length of the vector δn. Then, Eqn. (1.16) can be written
as
dφ
∆φn = ∇φ · ∆S n̂ =⇒ = ∇φ · n̂ (1.17)
¯ dS ¯
where dφ/dS is the rate of change of φ along n̂. Of special interest is the case when we find
encounter spatially constant fields. Let us say that that there exists a field φ(x, y, z) = constant
which represents a level surface. This means that if one moves along the surface of this field φ,
there would be no change in its value. Now, in Eqn. (1.16), let us choose δn to be a unit vector
that lies along the surface of this plane, i.e., in the tangential direction, t̂. Since φ is constant
along the horizontal surface, we can write Eqn. (1.16) as
∆φt = ∇φ · t̂ = 0 (1.18)
¯
This has to be valid for any t̂ chosen to lie in the plane of φ(x, y, z) = constant, i.e., ∇φ has to
¯
be normal to the surface φ(x, y, z) = constant. Physically, we are implying that the direction
of the greatest slope of a plane surface is normal to it! Does this make sense?
From Eqn. (1.15), we see that ∇ is a directional operator i.e., it’s components are associated
¯
with specific basis vectors. Further, the gradient operation of a scalar quantity, ∇φ yields a
¯
vector, i.e., there is no contraction of bases and the operation is similar to the dyadic product
19
outlined earlier. Thus, the gradient of a vector u would be represented as
∂ ∂ ∂
∇u= ê1 + ê2 + ê3 ⊗ (u1 ê1 + u2 ê2 + u3 ê3 )
¯ ∂x1 ∂x2 ∂x3
∂
= êi ⊗ uj êj
∂xi
∂uj
= êi ⊗ êj
∂xi
∂uj
=⇒ ∇ u = êi êj (1.19)
¯ ∂xi
The ∇ operator acts on a vector to result a second order tensor. We will see examples and
¯
applications of this later on in the course.
Divergence
In the previous section, we have seen that vectors can be involved in operations like dot
products, cross products and dyadic product. The del operator (∇ ) has been shown to be a
¯
vector and the gradient operation essentially forms a tensor product. We now focus on the scalar
or dot product using the del operator. Of course, a vector can be dotted only with another
vector or a tensor of higher order, but not a scalar. A simple example of a vector field is shown
in the figure below, which depicts the velocity variation in an idealized fluid flow (recall from
ME 202!). We recall that the general form of such a field is
and note that any component of the vector field can be dependent on all the three coordinates
at a given point.
div v = ∇ · v
¯
∂
= êi · vj êj
∂xi
∂vj
= êi · êj
∂xi
∂vj
= δij
∂xi
∂vi
=⇒ div v = ∇ · v = (1.21)
¯ ∂xi
20
The form of the above equation indicates that the divergence of v is a scalar, since no free
indices are present. Further, the divergence of v is a summation over the repeated index i,
given by
∂vi ∂v1 ∂v2 ∂v3
div v = = + + (1.22)
∂xi ∂x1 ∂x2 ∂x3
Physically, the divergence of a vector field describes and measures the propensity of a field
to radially converge or diverge at a point in space. Concomitant to the physical meaning of
divergence of a vector field, it can be quickly verified that for the simplistic example of fluid
flow presented earlier, the divergence of the field is 0 at all points in space. The implication of
the divergence operator is elucidated in greater detail in later sections.
Curl
The other vector algebraic operation is the vector or the cross product. When the del
operator (∇) acts on a vector field similar to the cross product, the result is called as the Curl
¯
of the vector field v. Recall that the vector product of two vectors also results in a vector. This
is given by
curl v = ∇ × v
¯
ê1 ê2 ê3
∂ ∂ ∂
=
∂x ∂x ∂x
1 2 3
v1 v2 v3
∂v3 ∂v2 ∂v3 ∂v1 ∂v2 ∂v1
= − ê1 − − ê2 + − ê3
∂x2 ∂x3 ∂x1 ∂x3 ∂x1 ∂x2
21
The partial derivative of any quantity with respect to xi is represented using ,i . Thus,
∂φ
grad φ = ∇φ = êi = φ,i êi
¯ ∂xi
∂uj
grad u = ∇ u = êi êj = uj,i êi êj
¯ ∂xi
(1.24)
∂ui
div u = ∇ · u = = ui,i
¯ ∂xi
∂uj
curl u = ∇ × u = ijk êk = ijk uj,i êk
¯ ∂xi
Integral theorems
The most commonly used integral theorem that relates integrals of fields is called as Gauss’
Theorem. However, Gauss’ theorem itself makes use of the result from the Green’s theorem
which relates a volume integral into an integral over its bounding surface. The Green’s theorem
states that, if f is a scalar field having continuous partial derivaties for each of the cartesian
basis êi , then
Z Z
∂f
dV = f ni dS
∂xi
V S
The theorem is also applicable in lower dimensions, i.e., an integral over a surface S can be
transformed to a contour C as Z Z
∂f
dS = f ni dC
∂xi
S C
From the above integrals, it is obvious that we can define f to be components of a vector
field v and the sum of the integrals over all components would yield
Z Z
∂vi
dV = vi ni dS
∂xi
V S
Z Z (1.25)
=⇒ ∇·v= v · n̂
¯
V S
The above result is called the Divergence theorem or Gauss’ theorem which states that
the divergence of a vector field over a volume is equal to the component of the vector along the
outward normal integrated over the surface that bounds the volume, as shown in Fig. 1.9. We
do have constraints that the vector field and its derivatives are continuous and that the surface
is closed.
The Gauss’ theorem can also be expressed in a more general form as
Z Z
∇ ? A dV = n̂ ? A dS (1.26)
¯
V S
where A could represent a scalar, vector or a tensor field and ? represents operations on the
fields such as the dot product, vector product or the dyadic product.
Another important integral transformation theorem relates the integral over a surface, S
to the integral over a closed contour, C, shown in Fig. 1.10. We do need to follow certain
22
Figure 1.9: Transformation of a volume integral into a surface integral - Gauss’ Theorem
conventions while defining the orientations. The tangent vector, t̂ along the contour C is
defined to conform to the right hand sense with the thumb pointing towards the direction of
the normal, n̂ of the surface S.
Figure 1.10: Transformation of a surface integral into a contour integral - Stokes’ Theorem
The general version of the Stokes’ theorem can then be written using A and ? defined above, as
Z Z
(n̂ × ∇) ? A dS = t̂ ? A dC (1.27)
¯
S C
For the specific case of A representing a vector field v and ? representing the dot product, we
get
Z Z
(∇ × v) · n̂ dS = t̂ · v dC (1.28)
¯
S C
23
Chapter 2
In the previous chapter, we have studied the mathematics and various notations that are neces-
sary for the description of continuous media. Beginning this chapter, we get into the application
of the ideas learnt so far to describe the phenomena associated with ‘continuous’ bodies. The
study of kinematics marks the first step in studying the mechanics of continuous media, the
other two being general laws governing continua and formation of constitutive relations for ma-
terials and phenomena. The study of kinematics enables us to quantify evolution of objects,
and correspondingly, fields, with time. In general, the emphasis is to study the ‘geometry of
motion’ without getting into details of the associated driving forces.
Let the collection of all the material points of a body be represented as B and let P be one
of the material points that belong to B. We use a coordinate system to represent the position
of any particle of an object. Let one such instantiation of B be represented by Ω and the
corresponding position of the point P be X. Note that X does not represent the particle
P . It only represents the position of the particle P in the configuration Ω. It follows that
24
Ω is basically one particular mapping of the body B. It is not necessary that the body B and
consequently, the point P are mapped only once. It is possible for B to have different mappings
at different instances of time. For the sake of continuity with respect to time, we choose one
such mapping to occur at time t = 0 and represent that mapping or that configuration as Ω0 .
Since we have imposed the constraint that each particle in B retains its uniqueness in each
configuration, this implies that the position vector X has a one to one mapping with P and is
invertible, i.e.,
X = Ω0 (P ) =⇒ P = Ω−1 0 (X)
x = Ωt (P ) = Ωt (Ω−1 −1
0 (X)) = Ωt ◦ Ω0 (X)
Note that the function/ mapping χ is represented as χ, i.e., its a vector valued function since
it yields the position vector x. We have, thus, outlined the method to describe the transition of
the body B through different states of space and time. Since we have considered Ω0 to represent
the configuration of B at t = 0, we usually refer to it as the reference configuration or as the
initial configuration or as the undeformed configuration. Ωt represents a subsequent state of B
and is referred to as the current configuration or as the deformed configuration. We can now
describe the motion of the particle P as
25
as it progresses through various stages of motion or deformation. Recall that we have chosen
Ω0 as the reference configuration in which each particle P belonging to the body B has been
identified with a unique position vector X, i.e., we already have a point-to-point map of each
particle P ∈ B onto Ω0 for which X acts as a parameter or a representative. Thus, following
or describing the motion of the particle P is anologous to following the change in X since X is
a unique mapping of P . As shown in Fig.2.1, the position of P at time t can be expressed in
terms of X as
Note that terming X as X(t) is redundant since we know t = 0 for the configuration Ω0 , but is
followed to retain uniformity in representation. We can write the velocity of the particle P as
the rate of change of position of P given by
∂V(X, t) ∂ 2 u(X, t)
A(X, t) = = (2.5)
∂t ∂t2
We point out that A is different from the rotation tensor A we have used earlier. From
here on, we use Q to represent rigid body rotation. We illustrate the validity of the Lagrangian
description of motion using a simple example - rigid body motion.
Consider the example shown in Fig. 2.2. We describe the rigid body motion of the body
from an initial configuration Ω0 to Ωt . Since we are citing rigid body motion, the displacement
is achieved via a rotation, specified by Q and a translation given by C. We emphasize there is
no change of shape (distortion) or change of volume (expansion) during rigid body motion, i.e.,
any two points have to remain the same distance apart during the motion. We consider two
points A and B, represented by position vectors XA and XB and the relative position vector
given by XAB = XB − XA . The distance between these two points in Ω0 is given by
q
X = XAB · XAB
AB
q
= XiAB êi · XjAB êj
q
= XiAB XjAB δij
q
= XiAB XiAB
The body is deformed from the Ω0 configuration to the Ωt configuration involving rigid body
rotation and translation. The mathematical mapping between X and x is given by
26
Figure 2.2: Rigid body motion using Lagrangian description.
Since there are no variations of deformation due to rigid body motion, C and Q are same
for all points of the body. Further, to simplify the equations, we drop the t in the expressions
since we know x = x(X, t). The relative position vector in the deformed configuration can be
written as
xAB = xB − xA
= (C + Q · XB ) − (C + Q · XA )
= Q · (XB − XA )
= Q · XAB
The magnitude of the relative position vector in the deformed configuration is given by
AB p AB AB
x = x · x
r
= Q · XAB · Q · XAB
q
= Qij XjAB êi · Qmn XnAB êm
q
= Qij XjAB Qmn XnAB δim
q
= XjAB Qij Qin XnAB
r T
= XjAB Qji Qin XnAB
q
= XjAB δjn XnAB
q
= XnAB XnAB
We have used the orthogonality of Q, i.e., QT · Q = I to verify that the Lagrangian descrip-
tion does indeed capture the kinematics of objects.
The other approach to describing the kinematics of motion is called as the Eulerian descrip-
tion of motion. Contrary to the Lagrangian approach where attention is fixed on a material
particle and its progression is tracked, the Eulerian approach places emphasis on a certain point/
window in space and tracks the movement of particles within that point/ domain of observation.
Thus, while we do observe material points pass through this ‘window’ and quantify the relevant
27
kinematic variables corresponding to those material points when they are in the window, we do
not follow them as we did in the Lagrangian approach. This means that the relevant kinematics
are described using parameters representing the point of observation, and not the particle itself.
Recall that in the Lagrangian approach, we had mapped each material point P to a position
vector X in Ω0 and followed that material point as it evolved to the Ωt configuration. If there is
a subsequent configuration at time, say 2t, we would still follow the same particle as it reaches
some new position x̄(X, 2t). In the Eulerian approach, the emphasis is on quantifying what
happens at the position x. At time t, this might correspond to the point P and at time 2t, it
might correspond to some other particle. However, we continue to observe x irrespective of the
material points which exists there at that instant of time. We do however, retain all information
of which particle is being mapped, i.e., χ(X, t) are defined and are used in the relevant analysis.
Since the mapping χ exists, the kinematics are directly derived from this as
28
Ω0 to deformed configurations, identified by Ω1 at time t = 1 and Ω2 at time t = 2. The initial
positions of the points are given by
XA
0 = 0 ê1 + 0 ê2
XB
0 = 1 ê1 + 0 ê2
XC
0 = 1 ê1 + 1 ê2
XD
0 = 0 ê1 + 1 ê2
If we desribe the motion of the body using a Lagrangian approach, the velocity and accleration
during the deformation can be given from Eqns. (2.4 and 5) as
∂χ(X, t)
V(X, t) = = 2X1 t ê1 + (X2 )2 ê2
∂t
The acceleration using the Lagrangian approach is given by
∂V (X, t)
A(X, t) = = 2X1 ê1
∂t
Let us choose material points B and C and follow the velocity and acceleration of these material
points respectively. The velocity of point B is determined to be VB (t = 1) = 2 ê1 and VB (t =
2) = 4ê1 . The acceleration of point C is determined to be AC (t = 1) = 2ê1 and AC (t = 2) = 2ê1 .
Now, we write the general expressions for the velocity and acceleration using the Eulerian
approach. The velocity is given by
∂χ(X, t)
v(x, t) = = 2X1 t ê1 + (X2 )2 ê2
∂t
Inspite of intending to describe v in terms of x, we still have v in terms of X. However, we
already know the relation between x and X, i.e.,
r
x1 x2
X1 = and X 2 =
1 + t2 1+t
29
using these relations, the velocity using the Eulerian approach can be given as
2tx1 x2
v(x, t) = ê1 + ê2
1 + t2 1+t
We know that x = 2 ê1 + 0 ê2 and t = 1 at the relevant point of observation. The velocity can be
determined to be vB (t = 1) = 2 ê1 and vB (t = 2) = 4 ê1 . So, we end up with the same velocity
that we have predicted previously, albiet with different representations, i.e., the velocity itself
and thus its magnitude are the same in both cases, but differ in their functional representations
since one is in terms of X and the other is in terms of x.
Now, let us determine the acceleration at the spatial point where the material point C is
present at any given instant of time. Using the expression for acceleration and the relation
between X and x, we can write
2x1
a(x, t) = 2X1 ê1 = ê1
1 + t2
At, t = 1 and x = 2 ê1 + 2 ê2 , aC (t = 1) = 2 ê1 and at t = 2 and x = 5 ê1 + 3 ê2 , aC (t = 2) = 2 ê1 .
We find out that the acceleration of a given material point is also the same, independent of the
approach used. However, it is important to understand the differences between the approaches.
In the first approach, we calculated the velocity of the material point B, VB (X, t) with the
knowledge that we were following point B. In the second approach, we were actually calculat-
ing the velocity of the material point present at coordinates (2, 0) at t = 1 and at coordinates
(5, 0) at t = 2. Of course, we chose the locations at those instances of time since we know that
they correspond to the material point B for a comparison to be made with the Lagrangian
approach. However, in general it is not necessary that we always follow the exact same material
point using both approaches.
For finding the acceleration using the Eulerian approach, we have simply used χ−1 to trans-
form from A to a. Let us verify rigorously if this actually holds true.
∂v(x, t)
a(x, t) =
∂t
∂ 2tx1 x2
= ê1 + ê2
∂t 1 + t2 1+t
−1
1 2t ∂x1
= 2 x1 + 2t 2tx1 + ê1
1 + t2 (1 + t2 )2 1 + t2 ∂t
−1
1 ∂x2
+ x2 + ê2
(1 + t)2 1 + t ∂t
Here, using the relations between x and X, we get
4t2
2x1 2 2t
a(x, t) = − X1 (1 + t ) + 2tX1 ê1
1 + t2 (1 + t2 )2 1 + t2
−1
2 (1 + t) + 1 2
+ X X ê2
(1 + t)2 2 1+t 2
2x1
= ê1
1 + t2
30
2.4 Material time derivative
In this chapter, we have seen methods for representing bodies/ objects in different states (config-
urations) as they progress through deformation and have addressed two approaches in describing
the deformation of a body. We have also seen that both approaches ultimately yield the same
description for a given material point, provided we have a clear relation between the differ-
ent configurations and mathematical relations between them, i.e., we have determined that
V(X, t) ≡ v(x, t) and A(X, t) ≡ a(x, t), the only difference between them being representation
in X or x. Inspite of the equivalence between the two representations, we have been dependent
on the mathematical relation between X and x via x = χ(X, t) and the inverse thereof. While
this might be a reasonable way to illustrate the idea of equivalence, it may not be feasible in
practical analysis to compute χ−1 at every point of interest. Further, such an approach would
imply that we would need to correlate the spatial point of observation x at time t and the
material point X at every instant of time (t) we make an observation. Thus, it would be more
operationally convenient and mathematically elegant if we could relate these two configurations
without explicit mathematical dependence on χ.
We recap here that when we compute the time derivative in the Lagrangian approach, we
are computing the time derivatives corresponding to a specific material point and since X is a
unique representation of that material point in the configuration Ω0 , the time derivative of that
material point is given by
∂V(X, t)
A(X, t) =
∂t
given X
However, when we find the acceleration using the Eulerian approach, we are capturing the
acceleration observed at the spatial point x, i.e.,
∂v(x, t)
a(x, t) =
∂t
given x
In the previous section, we related the spatial coordinate x with a specific material point using
χ−1 , i.e., we were calculating a(x, t) at the spatial point x that corresponds to the material
point X - we were still calculating the acceleration for a given X, but representing it in terms of
x. That is why we ended up with the same values for velocity and acceleration of a point using
both approaches - we were describing the same physical quantity! Thus, for our purpose here
of obtaining Lagrangian quantities (acceleration of a material point) using Eulerian approaches
(observing how things change at a given x), we will need to bear in mind that at the given
window of observation (x), the material point present at that instant of time is described by
χ(X, t), i.e., the change of the velocity of the material point located at x at the instant of time
t is given by
∂V(X, t) ∂v(x, t)
=
∂t ∂t
given X x=χ(X,t) for a given X
We note that v(x, t) is explicitly a function of t, but is also implicitly dependent on t via
31
x = χ(X, t). Thus, using the chain rule of differentiation, the above equation can be written as
Dv(x, t) ∂v(x, t)
= + ∇x v(x, t) · v(x, t) (2.9)
Dt ∂t ¯
This relation between the time derivative of a vector field associated with the material point
at x in terms of the time derivative of the vector field observed at x is not limited to the velocity
field alone. Indeed, the material time derivative of any vector field s can be written in terms of
the spatial time derivative of s as
Ds(x, t) ∂s(x, t)
= + ∇x s(x, t) · v(x, t) (2.10)
Dt ∂t ¯
32