You are on page 1of 33

Introduction to Continuum Mechanics

Lecture Notes

Jagan M. Padbidri

School of Engineering Sciences


Mahindra École Centrale

Fall 2017
Chapter 1

Essential mathematics

1.1 Vector algebra and indicial notation


Quantities that can be completely described by a single number are referred to as scalars. These
quantities are independent of any direction and physically, the number that denotes them refers
to the magnitude of the quantity. Typical examples for scalars are mass, temperature, density
etc. To describe quantities that are associated with a specific direction along with a certain
magnitude, vectors are used. Usually, one associates vectors with the description of physical
quantities such as force, velocity, acceleration etc.

Since our motivation is to describe physical phenomena using vectors, we will confine our-
selves to real valued vectors. The same holds true for scalars as well, i.e., we are interested in
real valued scalars. The set of all possible such scalars thus forms the real value space R. We
also define the set of all real valued vectors to form the space V. Both these spaces are linear
spaces and would need to satisfy certain conditions. Before listing the conditions, we will follow
that vectors will be designated using a bold face font or an underscore and unit vectors will be
identified with a carat (ˆ) symbol. Real valued vectors belonging to the linear vector space V
will need to satisfy the following properties.

Additive commutativity:
For any arbitrarily chosen vectors u, v ∈ V,
u+v=v+u
Here, it is also implied that the resultant sum is a vector and is also a member of the vector
space V. Thus, the addition of vectors is an operation that takes two elements of the vectors
space V and maps the resultant, also, on to the space V, i.e., this operation can be described
as V × V → V. The ‘×’ represents a generic operation, not mathematical multiplication.

Additive associativity:
For any arbitrarily chosen vectors u, v, w ∈ V,
(u + v) + w = u + (v + w)
This is also a V × V → V operation.

1
Zero element:
There exists an element 0 ∈ V such that
u+0=u

Negative element:
For every u ∈ V, there exists a v ∈ V such that
u + v = 0 =⇒ v = -u

Identity scalar multiplication:


For every u ∈ V,
1·u=u
Note that since 1 ∈ R , the above operation involves a scalar operating on a vector to result a
vector, i.e., an R × V → V operation.

Distributivity with vector addition:


For all α ∈ R and u, v ∈ V,
α (u + v) = αu + αv

Distributivity with scalar addition:


For all α, β ∈ R and u ∈ V,
(α + β) u = αu + βu

Associativity with scalar multiplication:


For all α, β ∈ R and u ∈ V,
(αβ) u = α(β u)

Scalar or Dot product


The dot product, or the scalar product of two vectors u, v ∈ V is given by
u · v = |u| |v| cos(θ)
where |u| and |v| are the magnitudes of the vectors u and v and θ is the angle between them.
Physically, the dot product of u and v represents the magnitude of v multipled with the
component of u in the direction of v. The converse is true as well. For the special case of
θ = 0, cos(θ) = 1, we have

u · u = |u||u| = |u|2

=⇒ |u| = u · u (1.1)

The dot product is an operation on two vectors that results in a scalar, i.e., it is a V × V → R
operation. Extending the possible values assumed by θ,
u · v = 0 =⇒ u ⊥ v (θ = π/2)

2
We note the dot product also satisfies the following conditions

u·v=v·u Commutativity
u · (v + w) = u · v + u · w Distributivity
α(u · v) = u · (αv) Associativity with scalars

Figure 1.1: (a)Dot product of two vectors and (b) Cross product of two vectors.

Vector or Cross product


The cross product or the vector product of two vectors, u and v results in another vector. The
magnitude of the cross product is given by |u × v | = |u| |v| sin(θ) where |u| and |v| are the
magnitudes of the vectors u and v and θ is the angle between them. The result of the cross
product is a vector w which is perpendicular to both u and v. Physically, the cross product of
u and v represents the area of the parallellogram formed by u and v.
|w| = u × v = |u| |v| sin(θ)
The orientation of the vector w is governed by the right hand rule. The cross product of two
vectors satisfies the following conditions

u × v = −v × u Anti-Commutativity
u × (v + w) = u × v + u × w Distributivity
α(u × v) = u × (αv) Associativity with scalars

Scalar triple product


The triple product or box product of three vectors u, v and w is given by
(u × v)·w = (v × w)·u = (w × u)·v
If the three vectors u, v and w form a right handed triad, then the triple product gives the
volume of the parallellopiped formed by them.

Figure 1.2: Schematic for scalar triple product.

Formation of a basis
The general rules for addition of vectors belonging to linear vector space can be extended to

3
an arbitrary set of vectors, i.e., the sum of an arbitrary set of vectors that belong to V should
result in a vector belonging to V. This can be written in a general form as follows. Given a set
of scalars (α1 , α2 , . . . , αn ) ∈ R and a set of vectors (u1 , u2 , . . . , un ) ∈ V,
α1 u1 + α2 u2 + · · · + αn un = v ∈ V

For the specific case of v = 0,


α1 u1 + α2 u2 + · · · + αn un = 0

If for an arbitrary set of vectors (u1 , u2 , . . . , un ) ∈ V, we can find a set of scalars (α1 , α2 , . . . , αn ) ∈
R, not all zero such that the above relation is satisfied, the set (u1 , u2 , . . . , un ) is said to be lin-
early dependent.

On the other hand, if


α1 u1 + α2 u2 + · · · + αn un = 0
implies that α1 = α2 = · · · = αn = 0, then the set (u1 , u2 , . . . , un ) is said to be linearly inde-
pendent.

Now, if we have a set of linearly independent vectors, (u1 , u2 , . . . , un ) and we chose a set of
scalars, (α1 , α2 , . . . , αn ), not all 0, then
α1 u1 + α2 u2 + · · · + αn un 6= 0
Rather,
α1 u1 + α2 u2 + · · · + αn un = v ∈ V

There are two important conclusions that can be drawn from this. First, since (α1 , α2 , . . . , αn )
are chosen arbitrarily and v ∈ V, this means that for any v ∈ V, a set of (α1 , α2 , . . . , αn ) can be
obtained to satisfy the above relation. Second, for a given v, the set (α1 , α2 , . . . , αn ) is unique
since (u1 , u2 , . . . , un ) is linearly independent. This means that any arbitrary v ∈ V can be
desribed using the linearly independent set (u1 , u2 , . . . , un ), i.e., (u1 , u2 , . . . , un ) acts as a basis
for the description of elements in V. Such a set of vectors are called the basis vectors for the
space V.

The most commonly used set of basis vectors are the ones used to describe vectors in
the cartesean coordinate system. While a vector space is not limited by dimension, since the
cartesean system describes 3D Euclidean space (R3 ), there are three basis vectors in this system.
We commonly refer to them as the x, y and z axes. Since these are basis vectors, any vector in
(R3 ) can be represented as a linear combination of these three basis vectors.

To simplify determining the set of scalars required to express an arbitrary vector in terms of
the basis vectors, we usually impose the condition that the basis vectors are of unit magnitude,
i.e., the vectors along the x, y and z directions are in fact x̂, ŷ and ẑ or î, ĵ and k̂. Also, the
vectors that form the basis are orthogonal to each other, i.e., î · ĵ = ĵ · k̂ = î · k̂ = 0. Hence,
this basis is called an Orthonormal basis. Now, any vector v ∈ R3 space can be expressed as a

4
Figure 1.3: Example of a basis in R3 .

linear combination of î, ĵ and k̂ as

v = vx î + vy ĵ + vz k̂ (1.2)

where vx , vy and vz are the components of v along î, ĵ and k̂ and represent the ‘weight’ of the
vector along said directions. It follows that

vx = v · î; vy = v · ĵ; vz = v · k̂ (1.3)

Figure 1.4: Example of representation of a vector v in a basis.

Note that it is not necessary that a vector space has a unique basis. It can have infinitely
many basis. However, the representation of any vector v ∈ R3 is unique for that specific basis.
We now revisit the vector operations previously addressed. Consider arbitrary vectors u and v
∈ R3 . The scalar product of the vectors can be written as

u · v = (ux î + uy ĵ + uz k̂) · (vx î + vy ĵ + vz k̂)


= ux î · vx î + ux î · vy ĵ + ux î · vz k̂
+ uy ĵ · vx î + uy ĵ · vy ĵ + uy ĵ · vz k̂
+ uz k̂ · vx î + uz k̂ · vy ĵ + uz k̂ · vz k̂
=⇒ u · v = ux vx + uy vy + uz vz (1.4)

Further, we can write the cross product of u and v as



î ĵ k̂

u × v = ux uy uz


vx vy vz

=⇒ u × v = (uy vz − uz vy )î − (ux vz − uz vx )ĵ + (ux vy − uy vx )k̂ (1.5)

5
Index notation
The above examples illustrate that even simple vector operations are rather cumbersome in rep-
resentation. We now introduce the notion of using indices to simplify representation of vectors
and avoid writing lengthy expressions. Since each vector is represented using the basis vectors,
we need to first address representing the basis vectors themselves in the indicial notation. We
replace the set of basis vectors (x̂, ŷ, ẑ) or (î, ĵ, k̂) with the set (ê1 , ê2 , ê3 ). This set can now be
generically represented as êi where i = 1, 2, 3. Note that the index i serves as a placeholder to
assume values 1, 2 or 3 and has no relation to î. We can express a vector as

v = vx î + vy ĵ + vz k̂
= v1 ê1 + v2 ê2 + v3 ê3
3
X
= vi êi
i=1

We further simplify the notation by introducing a summation convention: A repeated index


automatically implies summation over all possible values for that index. Using this convention,
the last term of the above equation automatically implies summation over all the possible
values for i even if summation is not explicitly specified. Given that we are concerned with 3D
Euclidean space, i would take values of 1, 2 and 3 implying

v = vi êi = v1 ê1 + v2 ê2 + v3 ê3 (1.6)

Note that the choice of using i as the index is arbitrary. Any other alphabet could be used as
an index, per convenience. We have already specified that the index i is not related to î. As
we have detailed previously, êi actually represents a set of three vectors (ê1 , ê2 , ê3 ) or (î, ĵ, k̂).
Similarly, êj represents the set of three vectors, albiet the exact number that i and j assume
can be arbitrary.

Now, consider the simple operation of finding the dot product of two unit vectors êi and
êj . We have already addressed the nature of the dot product when considering vectors from
an orthonormal triad. Just to summarize, êi and êj , each represent the relevant unit vector
depending on the value of the index and we are interested in finding êi · êj . Since êi and êj each
represent three vectors, we have a total of nine possibilities. They are

i = 1, j = 1; êi · êj = ê1 · ê1 = 1


i = 1, j = 2; êi · êj = ê1 · ê2 = 0
i = 1, j = 3; êi · êj = ê1 · ê3 = 0

i = 2, j = 1; êi · êj = ê2 · ê1 = 0


i = 2, j = 2; êi · êj = ê2 · ê2 = 1
i = 2, j = 3; êi · êj = ê2 · ê3 = 0

i = 3, j = 1; êi · êj = ê3 · ê1 = 0

6
i = 3, j = 2; êi · êj = ê3 · ê2 = 0
i = 3, j = 3; êi · êj = ê3 · ê3 = 1

The use of indicial notation allows us to account for all of these possibilities into the compact
form êi · êj , of course with the knowledge that i and j could take values between 1 and 3. We
can also make another, rather obvious observation from the above set of equations that

êi · êj = 1 if i = j
=0 otherwise

We simplify this by introducing a more convenient symbol called the Kronecker delta given
by

êi · êj = δij = 1 if i = j


=0 otherwise (1.7)

Note that δij has two indices i and j, each of which could take three possible values, i.e., δij
can account for all the nine possibilities shown above. Of course, of these possibilities, only
three have non-zero values (i = j = 1; i = j = 2; i = j = 3). The summation convention for the
repeated indices that we have previously introduced, holds good here as well i.e.,
δii = δ11 + δ22 + δ33 = 3

We note that the notion of finding the dot product of basis vectors can be used directly to
find the dot product of any two vectors. Since any arbitrary vector can be expressed in terms
of a set of scalar components and the basis vectors, we can use the properties of linear vector
spaces as,

u · v = ui êi · vj êj
= ui vj êi · êj
= ui vj δij

The above expression, ui vj δij implies that it is a summation over the repeated indices i and j.
Since both i and j can take values of 1, 2 or 3, the terms of the summation can be written as

ui vj δij = u1 v1 δ11 + u1 v2 δ12 + u1 v3 δ13


+ u2 v1 δ21 + u2 v2 δ22 + u2 v3 δ23
+ u3 v1 δ31 + u3 v2 δ32 + u3 v3 δ33
This can be rearranged as
ui vj δij = [u1 v1 δ11 + u2 v2 δ22 + u3 v3 δ33 ]
+ u1 v2 δ12 + u1 v3 δ13
+ u2 v1 δ21 + u2 v3 δ23
+ u3 v1 δ31 + u3 v2 δ32
All the δs outside the brackets reduce to 0
ui vj δij = [u1 v1 δ11 + u2 v2 δ22 + u3 v3 δ33 ] + 0

7
ui vj δij = u1 v1 + u2 v2 + u3 v3
=⇒ ui vj δij = ui vi
Comparing the respective terms, we get vj δij = vi

Alternatively, let us consider just the term vj δij . For any given value of the index i, vj δij
represents summation over three possible values of j, i.e., vj δij is the sum of three terms with
j taking on values 1, 2 and 3. In these three terms, there will be only one instance when
both i and j have the same value. For the remaining two terms, i 6= j =⇒ δij = 0 and
consequently, those two terms vanish! The only non-zero term in the summation corresponds
to when i = j =⇒ δij = 1 for that particular term. Thus, the summation vj δij condenses to
just one term, and i = j for that term and we can write

vj δij = vi (1.8)

Thus, δij acts on specific components of v and effectively ‘flips’ the index of that component by
rendering the remaining terms of the summation to 0. The dot product of two vectors can be
written as

u · v = ui vj δij
=⇒ u · v = ui vi (1.9)

Since the above expression has a repeated index, we can write out the summation over that
index and verify that the indicial notation does represent the dot product

u · v = ui vi = u1 v1 + u2 v2 + u3 v3

Now, let us consider the cross product of two vectors.

u × v = ui êi × vj êj = ui vj êi × êj

ui and vj are components of the vectors u and v i.e., they are scalars. The challenge lies in
describing the vector product êi × êj . This is seemingly more complex than describing the dot
product of two basis vectors since the result of a dot product is a scalar quantity, but we know
that the result of a cross product is a vector, subject to the right hand rule. Let us write down
the possibilities for êi × êj .

i = 1, j = 1; ê1 × ê1 = 0
i = 1, j = 2; ê1 × ê2 = ê3
i = 1, j = 3; ê1 × ê3 = −ê2

i = 2, j = 1; ê2 × ê1 = −ê3


i = 2, j = 2; ê2 × ê2 = 0
i = 2, j = 3; ê2 × ê3 = ê1

i = 3, j = 1; ê3 × ê1 = ê2

8
i = 3, j = 2; ê3 × ê2 = −ê1
i = 3, j = 3; ê3 × ê3 = 0

We observe certain patterns in the result of the vector product of two basis vectors. When
i = j, the vector product is 0. For the instances when i 6= j and the cross product is non-
zero, the indices i and j yield a vector which corresponds to the third index, k 6= i, j, i.e.,
ê2 × ê3 = ê1 , for example. Also, specific sequences of i and j produce a positive vector in
the right handed system and certain other sequences produce a negative vector. For example,
when i and j belong to the sequence 1, 2, 3, i.e., (i, j = 1, 2; i, j = 2, 3; i, j = 3, 1), the result
is a positive vector along êk , and when i and j are in the cyclically opposite sequence (3, 2,
1), i.e., (i, j = 2, 1; i, j = 3, 2; i, j = 1, 3), the result is a negative vector along êk . This can be
summarized as
êi × êj = ijk êk (1.10)
where ijk is called the permutation symbol whose value is given by

ijk =1 if i, j, k belong to 1, 2, 3 sequence,


=−1 if i, j, k belong to 3, 2, 1 sequence,
=0 otherwise.

We can now write the cross product of two vectors u and v as

w = u × v = ui êi × vj êj
= ui vj êi × êj
= ui vj ijk êk
=⇒ w = u × v = ijk ui vj êk (1.11)

1.2 Transformation of vectors, Introduction to tensors


Transformation of vectors

We have mentioned in the previous section that vectors have both magnitude and direction
and have seen the representation of vectors in an orthonormal triad of basis vectors. However,
we have not specifically addressed the uniqueness of this basis. It is rather obvious that the
three dimensional real Euclidean space (R3 ) we live in does not have a unique set of basis vectors
i.e., one can define a set of basis vectors per convenience and can also ‘switch’ between the bases
so defined. Note that we are referring to changing the basis for a given vector, not changing or
altering the vector itself. This is reflected as a change in the representation of the vector, but
not the vector itself. For example, if we have a vector v and two different bases êi and ēˆj , the
above argument can be summarized as

v = vi êi = v̄j ēˆj =⇒ vi 6= v̄j if êi 6= ēˆj

This still gives no information on relating vi and v̄i or even êi and ēˆi . We will address this by
formulating the transformation of a vector from one basis to another.

9
Consider a basis êi = (ê1 , ê2 , ê3 ) that has been rotated arbitrarily to obtain a new basis
ēˆi = (ēˆ1 , ēˆ2 , ēˆ3 ), as shown in Fig. 1.5. Note that both the bases are right handed orthonormal
triads. Since the basis vectors are not perfectly alligned with each other, each of the ‘old’
basis vectors, êi will have a non-zero component along each of the ‘new’ basis vectors, ēˆi . The
magnitudes of these components can be obtained from the dot products of the relevant basis
vectors, i.e., we can express each of the ‘new’ basis vectors as if they were an arbitrary vector in
the ‘old’ basis and vice versa, i.e., each of the ‘old’ basis vectors can be expressed as an arbitrary
vector in the ‘new’ basis.

ēˆ1 = (ēˆ1 · ê1 )ê1 + (ēˆ1 · ê2 )ê2 + (ēˆ1 · ê3 )ê3
= cos(ēˆ1 , ê1 )ê1 + cos(ēˆ1 , ê2 )ê2 + cos(ēˆ1 , ê3 )ê3
= cos θ11 ê1 + cos θ12 ê2 + cos θ13 ê3
=⇒ ēˆ1 = a11 ê1 + a12 ê2 + a13 ê3

Figure 1.5: Transformation of basis vectors.

It is emphasized here that one needs to be attentive to the conventions being used to describe
the angle between the ‘new’ and the ‘old’ basis. Here θ represents the angle between one ‘new’
basis vectors and one ‘old’ basis vector. The first index for θ represents the ‘new’ basis, ēˆi and
the second index represents the ‘old’ basis, êi , i.e., θ23 represents the angle between ēˆ2 and ê3 .
The other basis vectors in the ēˆi basis follow a similar mathematical pattern with respect to the
vectors belonging to the êi basis. Thus, the relations between the two bases can be expressed
as     
ēˆ1 a11 a12 a13 ê1
ˆ  
ē2  = a21 a22 a23  ê2 
 

ēˆ3 a31 a32 a33 ê3


or in the more succint indicial notation as ēˆi = aij êj . We denote the matrix shown above as
A, the transformation matrix. Similarly, we can express the ‘old’ basis vectors in terms of the
‘new’ ones as

ê1 = cos(ēˆ1 , ê1 )ēˆ1 + cos(ēˆ2 , ê1 )ēˆ2 + cos(ēˆ3 , ê1 )ēˆ3
= cos θ11 ēˆ1 + cos θ21 ēˆ2 + cos θ31 ēˆ3
ê1 = a11 ēˆ1 + a21 ēˆ2 + a31 ēˆ3
=⇒ êi = aji ēˆj

10
Now, we know that the ‘new’ basis vectors, ēˆi form an othonormal triad, i.e., ēˆi · ēˆj = δij . We
can write

ēˆi · ēˆj = aik êk · ajl êl


= aik ajl êk · êl
= aik ajl δkl
= aik ajk
δij = aik ajk
δij = aik (akj )T
=⇒ AAT = I

This yields us the very important observation that the transpose of the transformation matrix
is its inverse, i.e., the transformation matrix is orthogonal.

Now, consider a vector that is represented in both the bases - v̄ in the ēˆi basis and is
¯
equivalent to v in the êi basis. Using the previously obtained property of the transformation
matrix,
v̄ = v̄j ēˆj ≡ v = vi êi = vi aji ēˆj
¯
By comparing the relevant coefficients,

v̄j = aji vi =⇒ v̄ = Av = vAT


¯
The above equation prescribes the rules that need to be satisfied by entities to qualify them-
selves to be vectors, i.e., vector quantities have to transform from one basis to another described
by the relations above. We emphasize here that we are merely changing the basis for a given
vector, not altering the vector itself. From the above equations, we see that the components
of a vector change with a change in the basis in which it is being represented. Then, what do
we mean by the statement that the vector itself has not changed? As we shall see, there are
certain properties of vectors that are immune to the transformation from one basis to another.
In fact, the property of vectors to transform from one basis to another through a linear mapping
preserves the magnitude and the direction of the vector.

The idea that a vector has not changed or varied, i.e., remained invariant, implies that
the representation and method of quantification of a vector have not changed. For example,
consider the representation of the vector described above in both the bases, êi and ēˆi . The
√ √
magnitude of the vector in the basis êi is given by |v| = v · v = vi · vi and in the basis ēˆi
√ √
is |v̄| = v̄ · v̄ = v̄i · v̄i . It is straighforward to prove that the transformation of components
¯ ¯ ¯
between the two vectors, as derived above ensures that the magnitude is the same. However,
more fundamental is the expression for calculating the magnitude of the vector. We see that the
method/ expression to calculate the magnitude remains the same irrespective of the basis being
used i.e., the expression is invariant to the basis. This is the fundamental argument towards
the invariance of a vector with respect to the basis. The reader is advised to repeat the process
to explore the other quintessential property of a vector, its direction. Of course, since the basis

11
has been changed, so will the orientation of the vector described in that basis. Thus, the in-
variance of the direction of a vector needs to be verified using another arbitrarily defined vector.

We have discussed in detail the transformation of vectors from one basis to another. This
can be easily extended to describe rotation of vectors in the same basis. Instead of keeping our
vector fixed and rotating the basis vectors to a new position, we can keep the basis fixed and
define a ‘new’ vector by rotating the existing vector about the basis vectors. Determining the
components of this ‘new’ vector in the same basis is analogous to the previous case where the
vector was fixed and the basis was rotated. Thus, the transformation matrix defined above also
doubles up as a rotation matrix. Such an operation maps a vector from an ‘initial’ position in
the basis to a ‘final’ position/ state. Thus, the rotation matrix acts as a linear operator that
maps one vector into another.

Introduction to Tensors

The idea of mapping one vector to another using linear operators is common place, for
eg., the scalar mass is used to map the force and acceleration vectors. However, the spectrum
of mapping one vector to another has been left incomplete i.e., we have considered the more
convenient examples so far. Mapping the force and acceleration vectors takes place through a
physically measurable scalar and is very relavant physically, with both vectors oriented along
the same direction. We have discussed mapping one vector to another using the rotation matrix
with the directions of the two vectors being different, but the rotation matrix itself does not
correspond to any physically measurable quantity. We now introduce quantities which map/
transform vectors and have physical significance - Tensors.

Consider the following description of the angular momentum of a particle. Let the particle
have a mass m, position r = r1 ê1 + r2 ê2 and angular velocity ω = ω3 ê3 . We are familiar that
the angular momentum vector of the particle is given by h = Iω = mr×v, where v = ω×r is
the velocity of the particle. With a basic check, we can see that for r in the (ê1 , ê2 ) plane and
ω along ê3 , the velocity, v also lines in the (ê1 , ê2 ) plane. The angular momentum vector, h
= mr×v is oriented along ê3 . Thus, the role of the moment of inertia, I is analogous to mass
m in F = ma i.e., it maps ω to h with both vectors along the same direction. We note that
there exist significant differences between both cases since the force and acceleration vectors are
defined with respect to the origin of the basis, while angular momentum and angular velocity
can be defined about an arbitrary point or an axis that does not align with any individual basis
vector. Nevertheless, angular momentum and angular velocity are vectors which are mapped
onto each other by the moment of inertia.

Now, consider that the particle has the same angular velocity, ω = ω3 ê3 , but its position
vector is given by r = r1 ê1 + r2 ê2 + r3 ê3 . We can deduce that v = ω×r is still parallel to the
(ê1 , ê2 ) plane. However, the angular momentum vector, h = mr×v will not be along ê3 since r
does not lie in the (ê1 , ê2 ) plane. So here, the moment of inertia maps ω to h which are along
different directions. This deviates from our previous observation of I being analogous to m and

12
Figure 1.6: Schematic of a rotating mass and associated vectors.

one can deduce that I is not a mere scalar. Let us analyse the expression for angular momentum
in detail. We know,

h = mr × v
= mr × (ω × r)
= m(r · r)ω − m(r · ω)r

Written in the indicial notation, we get

h = hi êi = (mrj rj )ωi êi − (mωk rk )ri êi


= (mr1 2 + mr2 2 + mr3 2 )ω1 ê1 − (mω1 r1 + mω2 r2 + mω3 r3 )r1 ê1
+ (mr1 2 + mr2 2 + mr3 2 )ω2 ê2 − (mω1 r1 + mω2 r2 + mω3 r3 )r2 ê2
+ (mr1 2 + mr2 2 + mr3 2 )ω3 ê3 − (mω1 r1 + mω2 r2 + mω3 r3 )r3 ê3

= m(r2 2 + r3 2 )ω1 ê1 − m(r1 r2 )ω2 ê1 − m(r1 r3 )ω3 ê1


− m(r2 r1 )ω1 ê2 + m(r1 2 + r3 2 )ω2 ê2 − m(r2 r3 )ω3 ê2
− m(r3 r1 )ω1 ê3 − m(r3 r2 )ω2 ê3 + m(r1 2 + r2 2 )ω3 ê3

Similar to the rotation of a vector, the components of the angular momentum and angular
velocity vectors can be expressed as column vectors and the inertia in a matrix form as
       
h1 m(r2 2 + r3 2 ) −m(r1 r2 ) −m(r1 r3 ) ω1 I11 I12 I13 ω1
2 2
h2  =  −m(r2 r1 ) m(r1 + r3 ) −m(r2 r3 )  ω2  = I21 I22 I23  ω2 
       

h3 −m(r3 r1 ) −m(r3 r2 ) m(r1 2 + r2 2 ) ω3 I31 I32 I33 ω3

At first glance, the structure of the above equation looks very similar to that of transforma-
tion of vectors between bases. However, the inertia ‘matrix’, I and the transformation matrix,
A represent two fundamentally different quantities. The transformation matrix operates on a
vector to result its components in a new basis. Note that the composition of the transformation
matrix is independent of the vector it operates on i.e., the components aij are not derived from
the components of the vector they act on, but are derived from the two bases involved. The
inertia ‘matrix’, however, is explicitly derived from the position of the particle (r) and its mass.
Consequently, any change in r is summarily reflected in I. Thus, I has a more sound physical
basis and significance than A and represents a property/ state of the particle in the current

13
basis, in addition to performing the mathematical function of mapping ω to h. Such quanities
are called as Tensors. However, this does not imply that A is not a tensor. Indeed it is one
and belongs to a class of orthogonal tensors.

We have presented arguments that the components of the Inertia Tensor are derived from
components of the position vector, r. This implies that if one changes the basis such that the
components of r are altered, then the components of I should also change, i.e., the inertia tensor
(and tensors in general) are defined in a specific basis, similar to vectors.

Further we note that each component of I is associated with two basis vectors. For example,
consider how the second row of the I matrix operates. We have h2 = I21 ω1 + I22 ω2 + I23 ω3 . We
know that h2 represents the component of the angular momentum in the ê2 direction and ω3
represents the component of the angular velocity in the ê3 direction. The component I23 is thus
mapping a component along ê3 to a quantity along ê2 , i.e., each component of I is associated
with two basis vectors. It may be argued that I22 deals exclusively with the ê2 direction. The
relevant interpretation here is that I22 maps one component of the vector ω to one component
of the vector h - they just happen to be along the same direction! Thus, each component of the
tensor I is associated with two basis vectors and consequently, would require two basis vectors
to describe/ represent it.

Also evident is that a vector, which is associated with one basis vector for each component,
can be represented by a column, whereas I requires a two dimensional structure (matrix) for its
representation since each component of I is associated with two basis vectors. Thus, we refer
to I as a second order tensor. Similarly, vectors are referred to as first order tensors and scalars
are zero order tensors. Since we have identified inertia to be a second order tensor, we represent
it as I.

We have previously seen operations such as the dot product or the scalar product in which
a vector is reduced to a scalar, i.e., a first order tensor is reduced to a zero order tensor. Here,
we introduce the notion of tensor product or dyadic product represented by the ⊗ symbol. In
this operation, two first order tensors (vectors) are operated on to form a second order tensor
given by êi ⊗ êj = êi êj . The dyadic or tensor product of two vectors can be written as

u ⊗ v = ui êi ⊗ vj êj = ui vj êi ⊗ êj = ui vj êi êj = Tij êi êj

where Tij are the components of the second order tensor T formed by the tensor product of u
and v. The second order tensor has two free indices represented by the basis vectors êi and êj
and its components are represented as a matrix. In the present case, the the inertia tensor can
now be written as

Iij = m(rk rk δij − ri rj )


I = m[(r · r)I − r ⊗ r]

where I is the identity tensor given by I = ê1 ⊗ ê1 + ê2 ⊗ ê2 + ê3 ⊗ ê3 . Note that all tensors need
not be defined based on position vectors. There are many instances where second and higher or-

14
der tensors are used to describe a specific property of a material, independent of position vectors.

Once the Inertia tensor has been formed, it acts upon the angular velocity vector to result
the angular momentum vector. We have addressed the fact that the Inetia tensor is a second
order tensor (I) and the angular velocity vector is a first order tensor (ω). This implies that I
operates on ω to reduce the number of free indices i.e., the operation beween I and ω is a scalar
or dot product. This can be written as

h=I·ω
= Iij êi êj · ωk êk
= Iij ωk êi êj · êk
= Iij ωk êi δjk
=⇒ hi êi = Iij ωj êi

which is consistent with previous results.

Properties of Tensors

We have addressed the properties of vectors in a previous subsection in which we observed


that vectors belong to a linear space and since they are defined using a basis, they transform
according to certain rules. The same holds good for tensors as well. Tensors also belong to
the same space. Here, we distinguish between a space containing the collection of vectors, u, v
etc. and the vector space, V. The vector space, V that we have used thus far is not necessarily
limited to the collection of all possible, physical (R3 ) vectors. Rather, it is a more generalised
space of linear operators that satisfy the linear ‘vector’ space rules outlined in the first section.
Thus, all (R3 ) vectors belong to V, but do not completely define V. The tensors we have
defined can be viewed as linear operators on vectors, and so, belong to this general space of
linear operators and satisfy all the rules associated with belonging to a linear ‘vector’ space.
We have seen previously the rules governing change of vector representation with a change in
the basis and realized that quantities to be identified as vectors must transform according to
these specific rules. Since tensors are also expressed in a specific basis, there are specific rules
to be satisfied during transformation of bases. Consider the tensor I we have dealt with so far
in the êi basis and is transformed to Ī in the ēˆi basis.

I¯ij ēˆi ēˆj = Ikl êk êl


= Ikl amk ēˆm anl ēˆn
= Ikl amk anl ēˆm ēˆn
= amk Ikl (aln )T ēˆm ēˆn
=⇒ Ī = AIAT (1.12)

We note from the definition of the tensor I that when represented in a matrix, I is symmetric,
i.e., Iij = Iji . If, for some tensor A, Aij = −Aji then that tensor is said to be skew-symmetric.
To satisfy the requirements of skew-symmetry, the diagonal elements of such a tensor are 0.

15
In general, it is not necessary for a tensor to satisfy the requirements of symmetry or skew-
symmetry. However, any tensor can be represented as a sum of a symmetric and a skew-
symmetric tensor, i.e., any tensor C can be written as

C=A+B
!
C + CT
Aij = Aji is a symmetric tensor, where A =
2
!
C − CT
Bij = −Bji is a skew-symmetric tensor, where B =
2

We have seen that the Inertia tensor maps the angular velocity vector into the angular mo-
mentum vector. We summarize that the angular momentum is defined about a point that is
considered to be the origin and the angular velocity is a vector that passes through this origin.
The position vectors are also defined with respect to the origin. Also, the Inertia tensor maps
the angular velocity vector to the angular momentum vector such that they need not be along
the same direction. The formula for the angular momentum is given by

h = mr × v = m(r · r)ω − m(r · ω)r

The angular momentum vector, h is comprised of two different parts, one along ω and the
other along r. It is obvious that when ω ⊥ r, the part of h along r goes to 0 and hkω. We
can present these arguments about the angle between ω and r and the associated orientation of
h for the idealized example we have considered, i.e., a point mass which is described by a single r.

For a finite dimensional body, i.e., something which cannot be approximated to a point mass
and thus cannot be described by a unique position vector r, the above generalizations would not
hold. The Inertia tensor for such a body would have to be obtained by considering the moments
of inertia of infinitismal elements of the body and integrating them. For such an object, the
only mathematical possibility for h to lie along ω is for I to be a purely diagonal matrix, i.e.,
the off-diagonal terms should vanish. For each infinitismal elemet considered, the components
of the Inertia tensor of that element are functions of the components of the position vector of
that infinitismal element, i.e., I12 is a function of r1 and r2 , I13 is a function of r1 and r3 etc.
We recall that the components of any vector depend on the basis in which they are described.
Thus, one may select a basis to represent the finite dimensional body such that the off-diagonal
components of the Inertia tensor for a given infinitismal element may be non-zero, but when
that component of the inertia tensor is integrated over the domain of the body, it reduces to
zero. In such a case, the inertia tensor for the entire body would be represented by a diagonal
matrix and consequently, h k ω. Mathematically, this condition can be written as

h = I · ω = λω
I · ω − λω = 0
I · ω − λI · ω = 0
(I − λI) · ω = 0 (1.13)
We know that the solution for this exists when

16

I − λI = 0 (1.14)

The set of vectors which from the basis in which this condition is satisfied are called as Eigen-
vectors and the values of λ thus obtained are called as Eigenvalues. We can solve for these
by expressing the above equation in a matrix form as

I − λI = 0


=⇒ Iij − λIij = 0


I11 − λ 0 0

0 I22 − λ 0 =0


0 0 I33 − λ
which results in a cubic equation in λ i.e., three possible Eigenvalues, λI , λII and λIII . For each
of these Eigenvalues, there exixts a vector which satisfies Eqn. (1.13) and is the corresponding
Eigenvector, i.e., for λ = λI , there exists a vector ω I such that

(I − λI I) · ω I = 0

The set of vectors for all the three possible values of λ are called as the Eigenvectors. Physically,
this means that around this set of vectors/ axes the mass distribution is symmetric. For a tensor
whose components are real valued and symmetric, the Eigenvectors form an orthogonal triad.

1.3 Calculus of vectors and tensors


Thus far, we have addressed the algebra of vectors, which have been primarily defined at a
specific point at a specific instant of time. We have not considered cases of quantities varying
either spatially or temporally. In most physical situations, quantities do vary with both space
and time. Here we introduce the relevant mathematics to quantify these variations.

Consider a particle that moves with progression of time i.e., the position vector of the particle
changes with time and can thus be represented as a function of time as r = r(t). Let S be the
path that the particle traces. We consider the particle at two instances of time, t and t + ∆t.
Then the velocity of the particle is given by the derivate of the particle position as
∆r
v = lim
∆t→0 ∆t

If there are no discontinuities in the path S, then,



∆r
∆r
= ∆S
∆t ∆S ∆t

From the figure, it can be seen that as ∆t → 0, ∆r → ∆S and ∆r/∆S → êt where êt is the

tangential vector to the path S. Thus,


∆S dS
v = lim êt = êt
∆t→0 ∆t dt

17
Figure 1.7: Differentiation of the position vector.

i.e., the magnitude of the velocity vector is the rate at which distance is covered, but it is
oriented tangential to the path along which the particle is travelling. If we express êt in terms
of the basis vectors êi , then we can express the velocity as v = vi êi . Note that since êi s are
independent of time, t, this implies that the velocity can be integrated to obtain a general
expression for the position of the particle as r = ri (t)êi .

Gradient operator

Quite often in continuum description of physical properties and processes, we encounter


spatial variations of properties/ parameters, meaning, these quantities can be expressed as a
function of the coordinates of a given point in space, i.e., they can be expressed as Fields.
The simplest fields to imagine and to mathematically describe are scalar fields. Since the field
describes a scalar quantity, it is not associated with any particular direction. A typical example
could be the spatial distribution of temperature. The value of the temperature field at any
given point in space is a function of the components of the position vector of that point. The
general form of such a scalar valued function at a point P, is given by

φ = φ(x, y, z) = φ(x1 , x2 , x3 )

Note that the position vector of point P is given by P = xê1 + yê2 + zê3 ≡ x1 ê1 + x2 ê2 + x3 ê3 .
φ has been defined to be a continuous function dependent on all the three coordinates. Thus,
the total differential for φ is given by
∂φ ∂φ ∂φ
∆φ = δx + δy + δz
∂x ∂y ∂z
∂φ ∂φ ∂φ
= δx1 + δx2 + δx3
∂x1 ∂x2 ∂x3
∂φ
= δxi
∂xi
We note that the δxi s are essentially components of an infinitismal vector, δn, i.e., δn = δx1 ê1 +
δx2 ê2 + δx3 ê3 . We also note that the terms involving the partial derivatives could possibly
have directional dependence. We know that φ is a scalar valued function, i.e., it does not have
∂φ
components in specific directions. However, to evaluate , one finds the rate of change of
∂x1
φ only as x1 changes, which would obviously be along ê1 . Thus, the set of partial derivatives

18
evaluated are actually direction dependent and represent the rate of change of φ along the
directions of the basis vectors. The above equation for change of φ can then be written as
∂φ ∂φ ∂φ
∆φ = δx1 + δx2 + δx3
∂x1 ∂x2 ∂x3
 
∂φ ∂φ ∂φ
= ê1 + ê2 + ê3 · (δx1 ê1 + δx2 ê2 + δx3 ê3 )
∂x1 ∂x2 ∂x3

The function φ is a scalar function and does not have any components, implying that the
directional dependence in the above relation is derived from the operator determining the partial
derivative of φ. This operator is represented using the symbol ∇ and referred to as the Del or
¯
the grad or the nabla operator. Mathematically, it represents the Gradient of φ. Thus,
∂ ∂ ∂ ∂φ ∂φ ∂φ ∂φ
∇= ê1 + ê2 + ê3 such that ∇φ = ê1 + ê2 + ê3 = êi (1.15)
¯ ∂x1 ∂x2 ∂x3 ¯ ∂x1 ∂x2 ∂x3 ∂xi
The gradient of φ is rate at which the scalar function changes in space. The expression for
∆φ indicates that the total change in φ is equal to the rate of change of φ dotted with the
direction of interest. More specifically, the above equation would represent the change of φ
along a specific direction n given by

∆φn = ∇φ · δn (1.16)
¯
where δn has not been specified to be a unit vector. From the definition of the dot product, the
value of the the above dot product is maximum when δn k ∇φ. This means that ∆φ is maximum
¯
along the direction of ∇φ, i.e., the gradient of a field is a vector that gives the magnitude and
¯
direction of the greatest change for that field. The vector δn can be written in terms of a unit
vector as δn = ∆S n̂ where ∆S is the length of the vector δn. Then, Eqn. (1.16) can be written
as

∆φn = ∇φ · ∆S n̂ =⇒ = ∇φ · n̂ (1.17)
¯ dS ¯
where dφ/dS is the rate of change of φ along n̂. Of special interest is the case when we find
encounter spatially constant fields. Let us say that that there exists a field φ(x, y, z) = constant
which represents a level surface. This means that if one moves along the surface of this field φ,
there would be no change in its value. Now, in Eqn. (1.16), let us choose δn to be a unit vector
that lies along the surface of this plane, i.e., in the tangential direction, t̂. Since φ is constant
along the horizontal surface, we can write Eqn. (1.16) as

∆φt = ∇φ · t̂ = 0 (1.18)
¯
This has to be valid for any t̂ chosen to lie in the plane of φ(x, y, z) = constant, i.e., ∇φ has to
¯
be normal to the surface φ(x, y, z) = constant. Physically, we are implying that the direction
of the greatest slope of a plane surface is normal to it! Does this make sense?

From Eqn. (1.15), we see that ∇ is a directional operator i.e., it’s components are associated
¯
with specific basis vectors. Further, the gradient operation of a scalar quantity, ∇φ yields a
¯
vector, i.e., there is no contraction of bases and the operation is similar to the dyadic product

19
outlined earlier. Thus, the gradient of a vector u would be represented as
 
∂ ∂ ∂
∇u= ê1 + ê2 + ê3 ⊗ (u1 ê1 + u2 ê2 + u3 ê3 )
¯ ∂x1 ∂x2 ∂x3

= êi ⊗ uj êj
∂xi
∂uj
= êi ⊗ êj
∂xi
∂uj
=⇒ ∇ u = êi êj (1.19)
¯ ∂xi
The ∇ operator acts on a vector to result a second order tensor. We will see examples and
¯
applications of this later on in the course.

Divergence

In the previous section, we have seen that vectors can be involved in operations like dot
products, cross products and dyadic product. The del operator (∇ ) has been shown to be a
¯
vector and the gradient operation essentially forms a tensor product. We now focus on the scalar
or dot product using the del operator. Of course, a vector can be dotted only with another
vector or a tensor of higher order, but not a scalar. A simple example of a vector field is shown
in the figure below, which depicts the velocity variation in an idealized fluid flow (recall from
ME 202!). We recall that the general form of such a field is

v(x, y, z) = v1 (x, y, z)ê1 + v2 (x, y, z)ê2 + v3 (x, y, z)ê3 (1.20)

and note that any component of the vector field can be dependent on all the three coordinates
at a given point.

Figure 1.8: Example of the velocity field of fluid flow.

We define the Divergence of the vector field v, represented by div v as

div v = ∇ · v
¯
 

= êi · vj êj
∂xi
∂vj
= êi · êj
∂xi
∂vj
= δij
∂xi
∂vi
=⇒ div v = ∇ · v = (1.21)
¯ ∂xi

20
The form of the above equation indicates that the divergence of v is a scalar, since no free
indices are present. Further, the divergence of v is a summation over the repeated index i,
given by
∂vi ∂v1 ∂v2 ∂v3
div v = = + + (1.22)
∂xi ∂x1 ∂x2 ∂x3
Physically, the divergence of a vector field describes and measures the propensity of a field
to radially converge or diverge at a point in space. Concomitant to the physical meaning of
divergence of a vector field, it can be quickly verified that for the simplistic example of fluid
flow presented earlier, the divergence of the field is 0 at all points in space. The implication of
the divergence operator is elucidated in greater detail in later sections.

Curl

The other vector algebraic operation is the vector or the cross product. When the del
operator (∇) acts on a vector field similar to the cross product, the result is called as the Curl
¯
of the vector field v. Recall that the vector product of two vectors also results in a vector. This
is given by

curl v = ∇ × v
¯
ê1 ê2 ê3


∂ ∂ ∂
=
∂x ∂x ∂x

1 2 3

v1 v2 v3
     
∂v3 ∂v2 ∂v3 ∂v1 ∂v2 ∂v1
= − ê1 − − ê2 + − ê3
∂x2 ∂x3 ∂x1 ∂x3 ∂x1 ∂x2

This can be written in indicial form as


 

curl v = ∇ × v = êi × vj êj
¯ ∂xi
∂vj
= êi × êj
∂xi
∂vj
= ijk êk
∂xi
∂vj
=⇒ ∇ × v = ijk êk (1.23)
¯ ∂xi
The curl of a vector field is also referred to as the rotation of the vector field since it is propor-
tional to the angular velocity at a point in space. Again, the details are presented in subsequent
sections. A vector field for which ∇ × v = 0 at all points in space is called an irrotational
¯
field.

We introduce the following notation to simplify the representation of partial derivatives.

21
The partial derivative of any quantity with respect to xi is represented using ,i . Thus,
∂φ
grad φ = ∇φ = êi = φ,i êi
¯ ∂xi
∂uj
grad u = ∇ u = êi êj = uj,i êi êj
¯ ∂xi
(1.24)
∂ui
div u = ∇ · u = = ui,i
¯ ∂xi
∂uj
curl u = ∇ × u = ijk êk = ijk uj,i êk
¯ ∂xi
Integral theorems

The most commonly used integral theorem that relates integrals of fields is called as Gauss’
Theorem. However, Gauss’ theorem itself makes use of the result from the Green’s theorem
which relates a volume integral into an integral over its bounding surface. The Green’s theorem
states that, if f is a scalar field having continuous partial derivaties for each of the cartesian
basis êi , then
Z Z
∂f
dV = f ni dS
∂xi
V S
The theorem is also applicable in lower dimensions, i.e., an integral over a surface S can be
transformed to a contour C as Z Z
∂f
dS = f ni dC
∂xi
S C
From the above integrals, it is obvious that we can define f to be components of a vector
field v and the sum of the integrals over all components would yield
Z Z
∂vi
dV = vi ni dS
∂xi
V S
Z Z (1.25)
=⇒ ∇·v= v · n̂
¯
V S

The above result is called the Divergence theorem or Gauss’ theorem which states that
the divergence of a vector field over a volume is equal to the component of the vector along the
outward normal integrated over the surface that bounds the volume, as shown in Fig. 1.9. We
do have constraints that the vector field and its derivatives are continuous and that the surface
is closed.
The Gauss’ theorem can also be expressed in a more general form as
Z Z
∇ ? A dV = n̂ ? A dS (1.26)
¯
V S

where A could represent a scalar, vector or a tensor field and ? represents operations on the
fields such as the dot product, vector product or the dyadic product.

Another important integral transformation theorem relates the integral over a surface, S
to the integral over a closed contour, C, shown in Fig. 1.10. We do need to follow certain

22
Figure 1.9: Transformation of a volume integral into a surface integral - Gauss’ Theorem

conventions while defining the orientations. The tangent vector, t̂ along the contour C is
defined to conform to the right hand sense with the thumb pointing towards the direction of
the normal, n̂ of the surface S.

Figure 1.10: Transformation of a surface integral into a contour integral - Stokes’ Theorem

The general version of the Stokes’ theorem can then be written using A and ? defined above, as
Z Z
(n̂ × ∇) ? A dS = t̂ ? A dC (1.27)
¯
S C

For the specific case of A representing a vector field v and ? representing the dot product, we
get
Z Z
(∇ × v) · n̂ dS = t̂ · v dC (1.28)
¯
S C

23
Chapter 2

Kinematics of Continuous Media:


Configurations and motion

In the previous chapter, we have studied the mathematics and various notations that are neces-
sary for the description of continuous media. Beginning this chapter, we get into the application
of the ideas learnt so far to describe the phenomena associated with ‘continuous’ bodies. The
study of kinematics marks the first step in studying the mechanics of continuous media, the
other two being general laws governing continua and formation of constitutive relations for ma-
terials and phenomena. The study of kinematics enables us to quantify evolution of objects,
and correspondingly, fields, with time. In general, the emphasis is to study the ‘geometry of
motion’ without getting into details of the associated driving forces.

2.1 Configurations and motion


We define a continuous object as one that facilititates easy and unique description mathemat-
ically i.e., each point in the body should to be uniquely representable mathematically. This
implies that, for example, if two surfaces are in contact with one another, we represent the
contact as either being two points belonging to two surfaces/ bodies or that both surfaces have
combined to form a single object, but steer clear of representations of one physical point belong-
ing to two different objects. Thus, each point of an object has a unique position in space and
vice versa. This is a very fundamental requirement in developing the mathematics associated
with continuumm mechanics. When describing the mechanics of continua, we are interested in
capturing the physical phenomena that take place in the body. Contrary to our focus on spatial
representations thus far, we do not preclude the influence of time in descreptions henceforth,
i.e., the body itself and the processes associated with it are considered to be dependent on both
space and time.

Let the collection of all the material points of a body be represented as B and let P be one
of the material points that belong to B. We use a coordinate system to represent the position
of any particle of an object. Let one such instantiation of B be represented by Ω and the
corresponding position of the point P be X. Note that X does not represent the particle
P . It only represents the position of the particle P in the configuration Ω. It follows that

24
Ω is basically one particular mapping of the body B. It is not necessary that the body B and
consequently, the point P are mapped only once. It is possible for B to have different mappings
at different instances of time. For the sake of continuity with respect to time, we choose one
such mapping to occur at time t = 0 and represent that mapping or that configuration as Ω0 .
Since we have imposed the constraint that each particle in B retains its uniqueness in each
configuration, this implies that the position vector X has a one to one mapping with P and is
invertible, i.e.,
X = Ω0 (P ) =⇒ P = Ω−1 0 (X)

Similarly, we consider another mapping of B at time t, to the configuration Ωt . In this instance,


the position vector of the material point P is given by x. The uniqueness of mapping B to Ωt
follows and thus,
x = Ωt (P ) =⇒ P = Ω−1 t (x)

However, since P can be expressed in terms of X, we can write

x = Ωt (P ) = Ωt (Ω−1 −1
0 (X)) = Ωt ◦ Ω0 (X)

=⇒ x(t) = χ(X, t) (2.1)

Figure 2.1: Mapping of a body into configurations.

Note that the function/ mapping χ is represented as χ, i.e., its a vector valued function since
it yields the position vector x. We have, thus, outlined the method to describe the transition of
the body B through different states of space and time. Since we have considered Ω0 to represent
the configuration of B at t = 0, we usually refer to it as the reference configuration or as the
initial configuration or as the undeformed configuration. Ωt represents a subsequent state of B
and is referred to as the current configuration or as the deformed configuration. We can now
describe the motion of the particle P as

u = x(t) − X(t) (2.2)

2.2 Lagrangian and Eulerian descriptions of motion


The Lagrangian description of motion is intended to describe the motion or deformation of an
object from an individual particle point of view i.e., we aim to follow an individual particle

25
as it progresses through various stages of motion or deformation. Recall that we have chosen
Ω0 as the reference configuration in which each particle P belonging to the body B has been
identified with a unique position vector X, i.e., we already have a point-to-point map of each
particle P ∈ B onto Ω0 for which X acts as a parameter or a representative. Thus, following
or describing the motion of the particle P is anologous to following the change in X since X is
a unique mapping of P . As shown in Fig.2.1, the position of P at time t can be expressed in
terms of X as

x(X, t) = χ(X, t) = X(t) + u(X, t) (2.3)

Note that terming X as X(t) is redundant since we know t = 0 for the configuration Ω0 , but is
followed to retain uniformity in representation. We can write the velocity of the particle P as
the rate of change of position of P given by

∂x(X, t) ∂χ(X, t) ∂u(X, t)


V(X, t) = = = (2.4)
∂t ∂t ∂t
We are able to write V as V(X, t) since we know that we are describing the velocity of point
P , which is identified by X in Ω0 . Subsequently, the acceleration of P can be written as

∂V(X, t) ∂ 2 u(X, t)
A(X, t) = = (2.5)
∂t ∂t2
We point out that A is different from the rotation tensor A we have used earlier. From
here on, we use Q to represent rigid body rotation. We illustrate the validity of the Lagrangian
description of motion using a simple example - rigid body motion.

Consider the example shown in Fig. 2.2. We describe the rigid body motion of the body
from an initial configuration Ω0 to Ωt . Since we are citing rigid body motion, the displacement
is achieved via a rotation, specified by Q and a translation given by C. We emphasize there is
no change of shape (distortion) or change of volume (expansion) during rigid body motion, i.e.,
any two points have to remain the same distance apart during the motion. We consider two
points A and B, represented by position vectors XA and XB and the relative position vector
given by XAB = XB − XA . The distance between these two points in Ω0 is given by

q
X = XAB · XAB
AB
q
= XiAB êi · XjAB êj
q
= XiAB XjAB δij
q
= XiAB XiAB

The body is deformed from the Ω0 configuration to the Ωt configuration involving rigid body
rotation and translation. The mathematical mapping between X and x is given by

x(X, t) = χ(X, t) = C(X, t) + Q(X, t) · X(t)

26
Figure 2.2: Rigid body motion using Lagrangian description.

Since there are no variations of deformation due to rigid body motion, C and Q are same
for all points of the body. Further, to simplify the equations, we drop the t in the expressions
since we know x = x(X, t). The relative position vector in the deformed configuration can be
written as

xAB = xB − xA
= (C + Q · XB ) − (C + Q · XA )

= Q · (XB − XA )

= Q · XAB

The magnitude of the relative position vector in the deformed configuration is given by

AB p AB AB
x = x · x
r   
= Q · XAB · Q · XAB
q
= Qij XjAB êi · Qmn XnAB êm
q
= Qij XjAB Qmn XnAB δim
q
= XjAB Qij Qin XnAB
r   T
= XjAB Qji Qin XnAB
q
= XjAB δjn XnAB
q
= XnAB XnAB

We have used the orthogonality of Q, i.e., QT · Q = I to verify that the Lagrangian descrip-
tion does indeed capture the kinematics of objects.

The other approach to describing the kinematics of motion is called as the Eulerian descrip-
tion of motion. Contrary to the Lagrangian approach where attention is fixed on a material
particle and its progression is tracked, the Eulerian approach places emphasis on a certain point/
window in space and tracks the movement of particles within that point/ domain of observation.
Thus, while we do observe material points pass through this ‘window’ and quantify the relevant

27
kinematic variables corresponding to those material points when they are in the window, we do
not follow them as we did in the Lagrangian approach. This means that the relevant kinematics
are described using parameters representing the point of observation, and not the particle itself.
Recall that in the Lagrangian approach, we had mapped each material point P to a position
vector X in Ω0 and followed that material point as it evolved to the Ωt configuration. If there is
a subsequent configuration at time, say 2t, we would still follow the same particle as it reaches
some new position x̄(X, 2t). In the Eulerian approach, the emphasis is on quantifying what
happens at the position x. At time t, this might correspond to the point P and at time 2t, it
might correspond to some other particle. However, we continue to observe x irrespective of the
material points which exists there at that instant of time. We do however, retain all information
of which particle is being mapped, i.e., χ(X, t) are defined and are used in the relevant analysis.

Since the mapping χ exists, the kinematics are directly derived from this as

∂x(X, t) ∂χ(X, t) ∂u(X, t)


v(x, t) = = = (2.6)
∂t ∂t ∂t
for the particle with position x at the time t. The expressions for the velocity of a particle seem
exactly the same as they were for the velocity using the Lagrangian approach. This is because
the Eulerian approach measures the property of the particle at x at the time t. We know that
this is still, the particle P that we have followed using the Lagrangian approach. Obviously, a
material particle can have only one velocity, irrespective of the approach used to estimate it.
Similarly, the acceleration for the particle with position x at the time t can be given as

∂ 2 x(X, t) ∂ 2 χ(X, t) ∂ 2 u(X, t)


a(x, t) = 2
= 2
= (2.7)
∂t ∂t ∂t2
We do point out here that though we have argued that the values of the velocity or acceleration
in both approaches have to be the same, the functional form has to be different since the
Lagrangian approach is in terms of X and the Eulerian approach is in terms of x. Thus, to
establish equivalence between the two approaches, we use the inverse of Eqn. (2.1) in Eqns.
(2.6) and (2.7) as
 
∂χ χ −1 (x, t)
∂χ(X, t)
v(x, t) = =
∂t  ∂t  (2.8)
∂ 2 χ(X, t) ∂ 2 χ χ−1 (x, t)
a(x, t) = =
∂t2 ∂t2
We further detail the equivalence of both approaches using examples below.

2.3 Equivalence of Lagrangian and Eulerian descriptions


We present a simple example to compare the descriptions of motion using both Lagrangian and
Eulerian approaches. Consider the configurations shown in Fig. 2.3. We consider a simple case
of a 2D object. The body of interest is a square with corners described by points A, B, C and D.
We track the deformation of the set of particles as they deform from the initial configuration

28
Ω0 to deformed configurations, identified by Ω1 at time t = 1 and Ω2 at time t = 2. The initial
positions of the points are given by

XA
0 = 0 ê1 + 0 ê2

XB
0 = 1 ê1 + 0 ê2

XC
0 = 1 ê1 + 1 ê2

XD
0 = 0 ê1 + 1 ê2

Figure 2.3: Initial and deformed configurations to demonstrate equivalence.

The functions describing the deformation are given by


 
x(X, t) = χ(X, t) = X1 1 + t2 ê1 + (X2 )2 (1 + t) ê2

If we desribe the motion of the body using a Lagrangian approach, the velocity and accleration
during the deformation can be given from Eqns. (2.4 and 5) as

∂χ(X, t)
V(X, t) = = 2X1 t ê1 + (X2 )2 ê2
∂t
The acceleration using the Lagrangian approach is given by

∂V (X, t)
A(X, t) = = 2X1 ê1
∂t
Let us choose material points B and C and follow the velocity and acceleration of these material
points respectively. The velocity of point B is determined to be VB (t = 1) = 2 ê1 and VB (t =
2) = 4ê1 . The acceleration of point C is determined to be AC (t = 1) = 2ê1 and AC (t = 2) = 2ê1 .
Now, we write the general expressions for the velocity and acceleration using the Eulerian
approach. The velocity is given by

∂χ(X, t)
v(x, t) = = 2X1 t ê1 + (X2 )2 ê2
∂t
Inspite of intending to describe v in terms of x, we still have v in terms of X. However, we
already know the relation between x and X, i.e.,
r
x1 x2
X1 = and X 2 =
1 + t2 1+t

29
using these relations, the velocity using the Eulerian approach can be given as
2tx1 x2
v(x, t) = ê1 + ê2
1 + t2 1+t
We know that x = 2 ê1 + 0 ê2 and t = 1 at the relevant point of observation. The velocity can be
determined to be vB (t = 1) = 2 ê1 and vB (t = 2) = 4 ê1 . So, we end up with the same velocity
that we have predicted previously, albiet with different representations, i.e., the velocity itself
and thus its magnitude are the same in both cases, but differ in their functional representations
since one is in terms of X and the other is in terms of x.

Now, let us determine the acceleration at the spatial point where the material point C is
present at any given instant of time. Using the expression for acceleration and the relation
between X and x, we can write
2x1
a(x, t) = 2X1 ê1 = ê1
1 + t2

At, t = 1 and x = 2 ê1 + 2 ê2 , aC (t = 1) = 2 ê1 and at t = 2 and x = 5 ê1 + 3 ê2 , aC (t = 2) = 2 ê1 .
We find out that the acceleration of a given material point is also the same, independent of the
approach used. However, it is important to understand the differences between the approaches.
In the first approach, we calculated the velocity of the material point B, VB (X, t) with the
knowledge that we were following point B. In the second approach, we were actually calculat-
ing the velocity of the material point present at coordinates (2, 0) at t = 1 and at coordinates
(5, 0) at t = 2. Of course, we chose the locations at those instances of time since we know that
they correspond to the material point B for a comparison to be made with the Lagrangian
approach. However, in general it is not necessary that we always follow the exact same material
point using both approaches.

For finding the acceleration using the Eulerian approach, we have simply used χ−1 to trans-
form from A to a. Let us verify rigorously if this actually holds true.

∂v(x, t)
a(x, t) =
∂t
 
∂ 2tx1 x2
= ê1 + ê2
∂t 1 + t2 1+t
−1
 
1 2t ∂x1
= 2 x1 + 2t 2tx1 + ê1
1 + t2 (1 + t2 )2 1 + t2 ∂t
−1
 
1 ∂x2
+ x2 + ê2
(1 + t)2 1 + t ∂t
Here, using the relations between x and X, we get
4t2
 
2x1 2 2t
a(x, t) = − X1 (1 + t ) + 2tX1 ê1
1 + t2 (1 + t2 )2 1 + t2
−1
 
2 (1 + t) + 1 2
+ X X ê2
(1 + t)2 2 1+t 2
2x1
= ê1
1 + t2

30
2.4 Material time derivative
In this chapter, we have seen methods for representing bodies/ objects in different states (config-
urations) as they progress through deformation and have addressed two approaches in describing
the deformation of a body. We have also seen that both approaches ultimately yield the same
description for a given material point, provided we have a clear relation between the differ-
ent configurations and mathematical relations between them, i.e., we have determined that
V(X, t) ≡ v(x, t) and A(X, t) ≡ a(x, t), the only difference between them being representation
in X or x. Inspite of the equivalence between the two representations, we have been dependent
on the mathematical relation between X and x via x = χ(X, t) and the inverse thereof. While
this might be a reasonable way to illustrate the idea of equivalence, it may not be feasible in
practical analysis to compute χ−1 at every point of interest. Further, such an approach would
imply that we would need to correlate the spatial point of observation x at time t and the
material point X at every instant of time (t) we make an observation. Thus, it would be more
operationally convenient and mathematically elegant if we could relate these two configurations
without explicit mathematical dependence on χ.

We recap here that when we compute the time derivative in the Lagrangian approach, we
are computing the time derivatives corresponding to a specific material point and since X is a
unique representation of that material point in the configuration Ω0 , the time derivative of that
material point is given by
∂V(X, t) 
A(X, t) =

∂t

given X

However, when we find the acceleration using the Eulerian approach, we are capturing the
acceleration observed at the spatial point x, i.e.,

∂v(x, t) 
a(x, t) =

∂t

given x

In the previous section, we related the spatial coordinate x with a specific material point using
χ−1 , i.e., we were calculating a(x, t) at the spatial point x that corresponds to the material
point X - we were still calculating the acceleration for a given X, but representing it in terms of
x. That is why we ended up with the same values for velocity and acceleration of a point using
both approaches - we were describing the same physical quantity! Thus, for our purpose here
of obtaining Lagrangian quantities (acceleration of a material point) using Eulerian approaches
(observing how things change at a given x), we will need to bear in mind that at the given
window of observation (x), the material point present at that instant of time is described by
χ(X, t), i.e., the change of the velocity of the material point located at x at the instant of time
t is given by

∂V(X, t)  ∂v(x, t) 
=
 
∂t ∂t
 
given X x=χ(X,t) for a given X

We note that v(x, t) is explicitly a function of t, but is also implicitly dependent on t via

31
x = χ(X, t). Thus, using the chain rule of differentiation, the above equation can be written as

∂v(x, t)  ∂v(x, t)  ∂v(x, t) ∂x


= +
 
∂t ∂t ∂x ∂t
 
given X given x
∂v(x, t) 
  
= + ∇x v(x, t) · v(x, t)
∂t

given x ¯
where ∇ is the spatial gradient operator defined previously. We denote the relevant gradient
¯
here as ∇x since we are finding the gradient at the spatial location x. We note that the above
¯
equation gives the time derivative of the velocity of the material point present at x at the instant
of time, t. This is different from the time rate of change of v observed at x which is the first
term of the above equation. This term represents the rate at which the velocity of particles
observed to be at x changes and is thus not representative of the rate of change in velocity of the
material particle at x. The second term in the above equation compensates for this inadequacy
and gives a measure of the varition of velocity with space since the points neighbouring to x
need not have the same velocity. Since the above equation describes the Total derivative or
the derivative of the material point at x, it is referred to as the Material derivative and is

represented by D Dt, given by

Dv(x, t) ∂v(x, t)  
= + ∇x v(x, t) · v(x, t) (2.9)
Dt ∂t ¯
This relation between the time derivative of a vector field associated with the material point
at x in terms of the time derivative of the vector field observed at x is not limited to the velocity
field alone. Indeed, the material time derivative of any vector field s can be written in terms of
the spatial time derivative of s as

Ds(x, t) ∂s(x, t)  
= + ∇x s(x, t) · v(x, t) (2.10)
Dt ∂t ¯

32

You might also like