Professional Documents
Culture Documents
i
2. the associative law λ(µv) = (λµ)v holds
ora
3. the distributive laws (λ + µ)v = λv + µv and λ(u + v) = λu + λv hold
4. 1v = v
Gh
Definition: Elements v1 , v2 , · · · , vn of V are called linearly dependent if there exits
real number λ1 , λ1 , · · · , λn not all zero such that
λ 1 v 1 + λ2 v 2 + · · · + λn v n = 0
:S
independent.
Definition: A vector space V is said to be finite dimensional if there exists a finite
set of vectors v̂1 , v̂2 , · · · , v̂n such that
Dr
The set {v̂1 , v̂2 , · · · , v̂n } is called a basis of V and the above representation is unique.
λ1 , λ1 , · · · , λn are called coordiantes of v w.r.t. the basis {v̂1 , v̂2 , · · · , v̂n }.
Definition: A three dimensional vector space V is said to be Eucledian vector space
if for (u, v) ∈ V × V, we can define the following two operations u · v ∈ R and
(scaler product) u ∧ v ∈ V with the following properties:
1. u · v = v.u
2. (αu + βv) · w = αu · w + βv · w
2
3. u · u ≥ 0 and u · u = 0 iff u = 0
4. u ∧ v = −v ∧ u
6. u · (u ∧ v) = 0
7. (u ∧ v) · (u ∧ v) = (u · u)(v · v) − (u · v)2
i
|u| := (u · u)1/2
ora
The vectors u, v is called orthogonal if u · v = 0.
Definition: The scaler triple product of three vectors u, v, w is defined by
Gh
[u, v, w] = u · (v ∧ w)
Properties:
:S
written as
u = u1 e1 + u2 e2 + u3 e3
2 Indicial Notation
It is obvious that the index i in (2) can be replaced by j or m. The index i is (2) is
a dummy index in the sense that the sum is independent of the letter used.
We can further simplify the writing of (1) if we adopt the Einstein’s summation
convention: Whenever an index is repeated once, it is a dummy index indicating a
summation with index running through the natural numbers 1, 2, · · · , n.
Using this convention, (1) shortens to
i
ora
s = ai x i (3)
n = 3. Thus
ai x i = a1 x 1 + a2 x 2 + a3 x 3
Dr
The summation convention can be used to express double sum, triple sum, etc. For
example we can write the double sum involving 9 terms as
3 X
X 3
aij xi xj = aij xi xj
i=1 j=1
For beginers, it is better to perform the above expansion in two steps, first, sum
over i and then sum over j. Similarly the triple sum involving 27 terms is given by
3 X
X 3 X
3
aijk xi xj xk = aijk xi xj xk
i=1 j=1 k=1
As before the expression aii xi xj xj or aijk xi xj xk are not defined in the summation
convention.
4
x01 = a1m xm
i
ora
x02 = a2m xm
x03 = a3m xm
An index which appears only once in each term of an equation such as index i in
:S
(5) is called free index. A free index takes on values 1, 2 or 3 one at a time. Also
x0j = ajm xm , j = 1, 2, 3 is the same as (5). However
aft
ai = b j
is a meaningless equation. The free index appearing in every term of the equation
Dr
ai + b i = c i
ai + bi cj dj = 0
Tij = Tik
have no meaning.
5
1. δii = 3
2. δim am = ai
i
ora
3. δim Tmj = Tij
Also the the relations δim δmj = δij and δim δmj δjn = δin are easy to prove. If e1 , e2 , e3
are mutually orthonormal vectors then
Gh
ei · ej = δij
ijk = −1 if(ijk) form an odd permuation of (123) (8)
0 otherwise
Dr
Thus we have
ijk = jki = kij = −jik = −ikj = −kji
The permutation symbol allows us to write determinant of an 3 × 3 matrix in a very
compact form
T11 T12 T13
T21 T22 T23 = pqr T1p T2q T3r = pqr Tp1 Tq2 Tr3
T31 T32 T33
If {e1 , e2 , e3 } be an orthonormal basis for Eucledian vector space V, then the fol-
lowing relation holds
e2 ∧ e3 = ±e1
e3 ∧ e1 = ±e2
e1 ∧ e2 = ±e3 ,
6
ei ∧ ej = ±ijk ek
a · ej = ai ei · ej = ai δij = aj
i
{e1 , e2 , e3 } to be right hand triad. In this case [e1 , e2 , e3 ] = 1.
ora
Now we can write vector product in terms of components.
u ∧ v = (ui ei ) ∧ (vj ej ) = ui vj ei ∧ ej
Gh
= ijk ui vj ek
= (u2 v3 − u3 v2 )e1 + (u3 v1 − u1 v3 )e2 + (u1 v2 − u2 v1 )e3
v1 v2 v3
then
Tii = λθδii + 2µEii = 3λθ + 2µEii .
i
Definition: A second order tensor A is a linear transformation of three dimensional
ora
vector space V into itself. Thus
If A and B are two second order tensors, we can define the sum A + B, difference
A − B and scaler multiplication αA by the following rules
(A ± B)u := Au + Bu (10)
:S
There are two particularly important second order tensors (or simply tensors) ,
identity tensor I and zero tensor O which are defined by
Iu = u and Ou = 0
Dr
Using the definitions (10) and (11), we see that the set of all tensors (i.e. second order
tensors) is a linear vector space. We shall denote this space by CT (2). Subsequenly
we shall prove that this is finite dimensional and hence any tensor can be expressed
as a linear combination of simple tensors.
Definition: The tensor (or dyadic) product of two vector u and v denoted by u ⊗ v
and represents the linear transformation defined by
3. In general u ⊗ v 6= v ⊗ u
αij ei ⊗ ej = O, αij ∈ R
i
ora
Let p ∈ {1, 2, 3} be fixed but arbitrary. Then
A = Aij ei ⊗ ej , Aij ∈ R
Similarly
i
the components in the first column of the matrix are the components of the
ora
vector Ae1 etc. It should be noted that the components of A depend on the
coordinate system being used through the basis B = {e1 , e2 , e3 } just like the
components of a vector. A vector does not depend on any coordiante system
Gh
even though its components do. Similarly a tensor does not depend on any
coordinate system even though its components do.
(u⊗v)ij = up vq ei ·(ep ⊗eq )ej = up vq ei ·(ej ·eq )ep = up vq δjq ei ·ep = up vq δjq δip = ui vj
u1 v 1 u1 v 2 u1 v 3
u2 v 1 u2 v 2 u2 v 3
u3 v 1 u3 v 2 u3 v 3
I = ep ⊗ ep
Indeed, we have
(ep ⊗ ep )u = u = Iu
Thus the components of I are given by
1 0 0
0 1 0
0 0 1
10
u · (AT v) = v · (Au)
(AB)u = A(Bu), ∀u ∈ V
i
ora
4. A is said to be positive semi-definite if u · (Au) ≥ 0 ∀u ∈ V. And it is called
positive definite if u · (Au) > 0 ∀0 6= u ∈ V.
6. A is said to be orthogonal if
:S
AT A = AAT = I
aft
Observations:
ei · (Aej ) = ej · (AT ei )
Thus
Aij = ATji
Au = tA ∧ u, ∀u ∈ V
i
ora
= (A12 u2 + A13 u3 )e1 + (A23 u3 − A12 u1 )e2 + (−A13 u1 − A23 u2 )e3
Thus if tA = (tA A A
1 , t2 , t3 ), then
tA
1 = −A23 , tA
2 = A13 , tA
3 = −A12
Gh
Observation: If Au = tA ∧ u, ∀u ∈ V then A is skew-symmetric.
Proof: For any x, y ∈ V we have
:S
x · (Ay) = x · (tA ∧ y)
= −y · (tA ∧ x)
aft
= −y · (Ax)
Hence AT = −A
Dr
5. (u ⊗ v)T = v ⊗ u
Proof: Let A = (u ⊗ v). Then
Now
i
ora
T a = [(u ⊗ v)(w ⊗ x)]a = (u ⊗ v)[(w ⊗ x)a] = (u ⊗ v)[(a · x)w]
= (a · x)(v · w)u
and
Gh
Sa = (v · w)(u ⊗ x)a = (v · w)(a · x)u.
and
aft
Now
and
T a = [u ⊗ (AT v)]a = [a · (AT v)]u
Application: Let us find the components of the second order tensor T = AB. We
have A = Aij ei ⊗ ej and B = Bij ei ⊗ ej . Now
Thus we have
T = Tij ei ⊗ ej , with Tij = Aip Bpj
13
u = ui ei
Now choose a second orthonormal basis ê1 , ê2 , ê3 and we may write u in this new
i
ora
basis as
u = ui êi
Thus we have
êi · êj = Qij
:S
which are called the direction cosines of êi relative to ej . Now we have
aft
Thus [Q] is an orthogonal matrix. Now we have det[Q] = ±1. If det[Q] = 1 then
[Q] is a proper orthogonal matrix representing rotation. If det[Q] = −1, then [Q] is
improper orthogonal which represents a rotation combine with a reflection.
We now have the inverse transformation as
Also since
u = uj ej = uj Qij êi , u = ûi êi
we get
ûi = Qij uj
and similarly
ui = Qji eˆj
i
ora
Next we consider the transformation law for tensors. We have
A = Aij ei ⊗ ej ,
Gh A = Âij êi ⊗ êj
Now
Aij = ei · Aej , and Âij = êi · Aeˆj
Thus
Âij = (Qip ep ) · A(Qjq eq ) = Qip Apq Qjq
:S
and similarly
aft
under a change of basis to B̂ = {ê1 , ê2 , ê3 } given by êi = Qij ej with [Q] a proper
3 × 3 orthogonal matrix. The set of all cartesian tensor of order n is a linear vector
space denoted by CT (n).
Note: In the definition of tensors using the alternated definition, we insist on proper
orthogonal transfomration. Since there are pseudo tensors which transforms accord-
ing to
Âi1 i2 i3 ···in = det(Q)Qi1 j1 Qi2 j2 · · · Qin jn Aj1 j2 j3 ···jn
15
i
ora
Definition: If the components of A ∈ CT (n) are unchanged under arbitrary rota-
tions of orthonormal basis, then A is said to be isotropic.
Examples:
Gh
1. The set of all scalers CT [0] are isotropic.
2. The zero vector 0 is the only isotropic vector. To satisfy the definition of
isotropy, we must have
:S
ûi = Qij uj = ui
for all rotations Qij . Let us choose
aft
0 1 0
[Qij ] = −1 0 0
Dr
0 0 1
3. Scaler multiple of δij are the only isotropic tensor of CT (2). We have
for all rotations [Q]. The choice of [Q] of the previous examples gives
Now taking
1 0 0
[Qij ] = 0 0 1
0 −1 0
16
gives
A22 = A23 , A21 = A31 = 0
i
ora
δij δkl , δik δjl , δil δjk
then
The third relations follows from the fact that if any two i, j, k are equal, then the
sum is zero, etc. Now
Again
i
ora
= [Ai1 ei , Aj2 ej , e3 ] + [Ai1 ei , e2 , Ak3 ek ] + [e1 , Aj2 ej , Ak3 ek ]
= Ai1 Aj2 [ei , ej , e3 ] + Ai1 Ak3 [ei , e2 , ek ] + Aj2 Ak3 [e1 , ej , ek ]
= (A11 A22 − A12 A21 ) [e1 , e2 , e3 ] + (A22 A33 − A23 A32 )[e1 , e2 , e3 ]
Gh
+ (A11 A33 − A13 A31 )[e1 , e2 , e3 ]
Thus
A
11 A12
A
22 A23
A
11 A13
1
IA2 = (trA)2 − tr(A2 )
+ + =
:S
Lastly, we have
aft
tr(AB) = tr(BA)
tr(AT ) = tr(A)
18
det(αA) = α3 det(A)
det(AB) = det(A)det(B)
det(I) = 1
det(A−1 ) = (det(A))−1
det(AT ) = det(A)
The invariants of IA1 , IA2 , IA3 can be proved by easily by calculating these components
i
ora
in different basis. For example we can prove that
A0ii = Ajj
Gh
where the left hand side is with respect to basis {e01 , e02 , e03 } and the right hand side
is with respect to {e1 , e2 , e3 }.
Definition: A scaler λ is an eigen value of A ∈ CT (2) if there exists a non-zero
vector p which transforms into a vector parallel to inself. Thus
:S
Ap = λp, or (A − λI)p = 0
aft
det(A − λI) = 0
(Note the above relation can be proved other way also. Take q and r such that
[p, q, r] 6= 0. Using A − λI for A and p, q and r for u, v and w we also get
det(A − λI) = 0.)
Now we can directly carry out the expansion of the determinant or employ the
invariance properties to expand the determinant.
Thus we have
Since [u, v, w] is arbitrary we obtain, after using the invariant properties and re-
moving out the factor [u, v, w], the characteristic equation for A as
Properties:
p · Ap = λp · p
i
ora
Since both p · Ap and p · p are positive, we conclude that λ > 0
2. If A is symmetric, then its charateristic equation has three real eigen values.
Proof: Since the characteristic equation is of degree three and of real coeffi-
Gh
cients, one of the eigen value is real and the corresponding eigen vector is real
too. Let n1 corresponds to the eigen values λ1 . Now we choose n2 and n3 such
that {n1 , n2 , n3 } is a right handed triad. If we shift to the new coordinate
system, then we have
:S
λ1 0 0
aft
[A]n1 ,n2 ,n3 =
0 a022 a023
0 0
0 a23 a33
The eigen values remains same and hence are given by λ = λ1 and the remaing
Dr
Due to symmetry, the left hand side is zero and thus since λ 6= µ, we have
u · v = 0.
i
A = λ1 n1 ⊗ n1 + λ2 n2 ⊗ n2 + λ3 n3 ⊗ n3
ora
The orthonormal basis {n1 , n2 , n3 } and corresponding eigen values λ1 , λ2 , λ3 are
called principal directions and principal values respectively.
Proof: We have already proved that if the eigen values λ1 , λ2 , λ3 are distinct, then
Gh
the corresponding eigen values u, v, w are mutually orthogonal. We normalize the
eigen vectors to unit length and then let n1 = u, n2 = v, n3 = w. Thus
λ1 0 0
:S
[A]n1 ,n2 ,n3 =
0 λ 2 0
0 0 λ3
aft
Now suppose that λ1 6= λ2 = λ3 are the eigen values. Let n1 corresponds to the
eigen values λ1 . Now we choose n2 and n3 such that {n1 , n2 , n3 } is a right handed
triad. If we shift to the new coordinate system, then we have
Dr
λ1 0 0
[A]n1 ,n2 ,n3 = 0 0
0 A 22 A 23
0 A23 A033
0
Since the tensor has eigen values λ1 , λ2 = λ3 we must have A022 = A033 = λ2 and
A023 = 0. Thus
λ1 0 0
[A]n1 ,n2 ,n3 =
0 λ2 0
0 0 λ3
In this case any vector in the plane of n2 , n3 is an eigen vector.
(Proof: Let u = αn2 + βn3 . Now Au = αAn2 + βAn3 . Now let An2 = γ1 n1 +
γ2 n2 + γ3 n3 . Since n1 · An2 = n2 · An1 = λ1 n1 · n2 = 0 due to symmetry we have
γ1 = 0. Also n3 · An2 = A023 = 0 and hence γ3 = 0 etc. Thus γ1 = λ2 etc.)
21
i
ora
A = λ1 n1 ⊗ n1 + λ2 n2 ⊗ n2 + λ3 n3 ⊗ n3
Hence
A−1 = λ−1 −1 −1
1 n1 ⊗ n1 + λ2 n2 ⊗ n2 + λ3 n3 ⊗ n3
Gh
b. The principal values of a tensor A contains the maximum and minimum values
that the diagonal elements of any matrix of A can have. Consider any unit
vector n = αn1 + βn2 + γn3 . Now
:S
Ann = n · An = λ1 α2 + λ2 β 2 + λ3 γ 2
we have
λ1 = λ1 (α2 + β 2 + γ 2 ) ≥ λ1 α2 + λ2 β 2 + λ3 γ 2 = Tnn
Dr
Similarly
Tnn = λ1 α2 + λ2 β 2 + λ3 γ 2 ≥ λ3 (α2 + β 2 + γ 2 ) = λ3
Now
f (A)x = αf (λ1 )u + βf (λ2 )v + γf (λ3 )w = ~0
Hence by definition
f (A) = 0
22
Square Root Theorem: Let A ∈ CT (2) be a symmetric positive tensor. Then there
is unique positive definite symmetric tensor B ∈ CT (2) such that B 2 = A.
Proof: (Existence) Since A is symmetric we may write
A = λ1 n1 ⊗ n1 + λ2 n2 ⊗ n2 + λ3 n3 ⊗ n3 ,
where λi are the eigen values of A. Since A is positive definite, λi > 0. Define B
by
√ √ √
B= λ1 n1 ⊗ n1 + λ2 n2 ⊗ n2 + λ3 n3 ⊗ n3 ,
Then
i
√ √ √
BT =
ora
λ1 n1 ⊗ n1 + λ2 n2 ⊗ n2 + λ3 n3 ⊗ n3 = B
3
X
= λi ni ⊗ ni = A
i=1
aft
B2 = C 2 = A
Dr
Let u an eigen vector of A with λ > 0 the corresponding eigen value. Then letting
√
β = λ,
0 = (B 2 − λI)u = (B + βI)(B − βI)u
Let
v = (B − βI)u,
then
Bv = −βv
Bu = βu
23
and similarly
Cu = βu
Thus Bu = Cu for every eigen vector v of A. Since we can form a basis of eigen
vectors of A, B and C must concide.
Theorem: A symmetric tesnor A is positive definite iff each of its principal values
is strictly positive.
Proof: Let A is positive definite and λ and u denote a principal value and direction
respectively. Now since u is unit vector we have
i
ora
due to A being positive deinite.
Next let the principal values are positive. Let n1 , n2 , n3 denote the principal
axes. Now we have
Gh
A = λ1 (n1 ⊗ n1 ) + λ2 (n2 ⊗ n2 ) + λ3 (n3 ⊗ n3 )
Now
:S
which is greater than or equal to zero. Let (u, Au) = 0, then uni = 0 and hence
u=0
Polar Decomposition Theorem: An arbitrary invertible tensor A ∈ CT (2) can be
expressed uniquely in the forms
A = RU = V R
U 2 = AT A, V 2 = AAT
24
RT R = I
Since det(U ) > 0, both the det(A) and det(R) have the same sign. Since U is
unique, R must be unique.
Proceeding in the same way we find the A = V P where P = V −1 A is orthogonal
and V is unique. Next we show that P = R. We have
A = RU = IA = (P P T )V P = P (P T V P )
i
Now P T V P is symmetric. Let u be an arbitray vector in V. Now
ora
u · P T V P = (P u) · V (P u) > 0
Double contraction: This is denoted by doble dots and for second order tensors it
results in a scaler and is defined by
Dr
A : B = tr(AT B)
Note that this operation is commutative. The norm of a tensor A denoted by |A|
is defined by
|A| = (A : A)1/2 = (Aij Aij )1/2 ≥ 0
a. I : A = tr(A) = A : I
b. A : (BC) = (B T A) : C = (AC T ) : B
c. A : (u ⊗ v) = u · Av = (u ⊗ v) : A
d. (u ⊗ v) : (w ⊗ x) = (u · w)(v · x)
25
4.1 Differentiation
In this section we introduce a notion of differentiation sufficiently general to include
scaler, point, vector, or tensor functions whose arguments are scalars, points, vectors
or tensors. Let E denote the point space. A point space can be associated with a
vector space V since the difference of two points is a vector. Also the sum of a point
and a vector is a point. We can associate norm with R, V and L(V), where L(V) is
the set of second order tensors.
i
Let U and W denote such normed vector spaces and consider g : U → W. Let
ora
g be defined in a neighbourhood of zero in U and have values in W. We say that
g(u) approaches zero faster than u, and write
Gh
g(u) = o(u) as u → 0
if
||g(u)||
lim =0
u→0,u6=0 ||u||
:S
Similarly
f (u) = g(u) + o(u)
aft
implies that
f (u) − g(u) = o(u)
Dr
The last definition makes sense if f and g have values in E, in that case f − g has
values in the vector space V.
As an example let φ(t) = tα . Then φ(t) = o(t) iff α > 1.
Let g be a function whose values are scalers, points, vectors, or tensors and
whose domain is an open interval D of R. The derivative ġ(t) of g at t, if it exists,
is defined by
d 1
ġ(t) = g(t) = lim [g(t + α) − g(t)]
dt α→0 α
or equivalently
g(t + α) = g(t) + αġ(t) + o(α)
derivative is based on this result: we define the derivative to be the linear map
which approximates g(t + α) − g(t) for small α.
Thus let U and W be finite dimensional normed vector spaces and let D be a
domain in U. Consider
g:U →W
g(x + u) − g(x)
is equal to a linear function of u plus a term that approaches zero faster than u.
i
ora
Thus g is differentiable at x if there exits a linear transformation
Dg(x) : U → W (19)
such that
Gh
g(x + u) = g(x) + Dg(x)[u] + o(u), as u → 0
Dv(x)[u] = ∇v(x)u
i
ora
div v = tr∇v
The curl of v, denoted by curl v, is the unique vector field with the property
aft
Thus curl v is the axial vector corresponding to the skew tensor ∇v − (∇v)T
Example: Let Φ be a smooth scaler, vector, or tensor field. Then
Dr
1
DΦ(x)[ei ] = lim [Φ(x + αei ) − Φ(x)]
α→0 α
∂Φ(x)
DΦ(x)[ei ] =
∂xi
From this we deduce
∂φ
∇φ = (∇φ)i ei = ei := φ,i ei
∂xi
(∇u) = ui,j ei ⊗ ej
∇ · u = ui,i
and
(div A) = Aij,i ej
v · n dS = ∇ · v dv
Z ∂D ZD
AT n dS = ∇ · A dv
i
∂D D
ora
Let us prove the third relation. Taking dot product with a constnt vector a we get
Z Z Z Z Z
T T
a· A n dS = a·A n dS == Aa·n dS = ∇·(Aa) dv = a· ∇·A dv
∂D ∂D
Gh ∂D D D
over an open surface to a line integral around the bounding bounding curve of the
surface. Le C be the bounding curve to the open surface S and dx be the tangential
aft
vector to C. If n̂ is the unit outward normal and v is any vector field defined on S
and C, Stoke’s theorem asserts that
Dr
Z Z
n̂ · (∇ ∧ v) dS = v · dx
S C