You are on page 1of 4

Linear Algebra - MATH 232 Cheat Sheet

by fionaw via cheatography.com/124375/cs/23750/

Basic Equations Vectors, Norm, Dot Product (cont) Matrix Algebra, Identity and Inverse
Matrix
Network Flows u·v = u1v1 + u2v2 dot product
1. the flow in an arc is only in one directions ...+unvn (A + B)ij = (A)ij + (B)ij (A - B)ij = (A)ij -
||u|| ||v|| cos(θ) (B)ij
2. flow into a node = flow out of a node
u and v are orthogonal if u·v = 0 (cos(θ) = 0) (cA)ij = c(A)ij (A T)ij = (A)ji
3. flow into the network = flow out of the
network a set of vectors is an orthogonal set iff vi·vj (AB)ij = ai1b1j + ai2b2j + ... aikbkj
= 0,if i≠j Inner Product (number) is u Tv = u·v, u
Bala​ncing Chemical Equati​ons
a set of vectors is an orthon​ormal set iff vi·vj and v same size
1. add x's before each combo and both side
= 0,if i≠j, and ||vi|| = 1 for all i
Outer Product (matrix) is uvT, u and v can
2. carbo = x1 + 2(x3), set as system, solve
(u·v)2 ≤ ||u||​2|​|v|​|2 Cauchy​-Sc​hwarz be any size
Matrix
or Inequality
(AT) T = A (kA)T = k(A)T
augmented variables and soluti​‐ |u·v| ≤ ||u|| ||v||
matrix on(rhs) (A+B)T = AT + BT (AB)T = BTAT
d(uv) ≤ d(u,w) + Triangle Inequality
coeffi​cient coeffi​cients only, no rhs tr(A T) = tr(A) tr(AB) = tr(BA)
d(w,v)
matrix ||u+v|| ≤ ||u|| + ||v|| uTv = tr(uvT) tr(uvT) = tr(vuT)

||v1 + v2 ... + vk|| = ||v1|| + ||v2|| ... + ||vk|| tr(A) = a11 + a22 ... + (AT)ij = Aji
Vectors, Norm, Dot Product
ann
maginitude (norm) of vector v is ||v||; ||v|| ≥ 0 Lines and Planes Identity matrix is square matrix with 1 along
if k>0, kv same direction magnitude = a vector equation with x = x0 + tv, diagonals
as v k||v|| parameter t -∞ < t < +∞ If A is m x n, A n = A and mA = A
if k<0, kv opposite magnitude = |k| solutin set for 3 dimension linear equation is a square matrix is AB = = BA
direction to v ||v|| a plane invertible(nonsingular)
vectors in Rn (n = v = (v1, v2, ..., if x is a point on this plane n·(x-x0) = 0 if:
dimension) vn) (point-normal equati​on) B is the inverse of A B = A-1
v = P1P2 = OP2 - OP1 displa​cement A(x-​x0)​+B(​y-y​0)+​C(z-z0) = x0 = if A has no inverse, A is not invertible
vector 0 (x0,y0​,z0), (singular)
norm/m​agn​itude of sqrt( (v1)2​+(​‐ n = (A, B, C)
det(A) = ad - bc ≠ 0 is invert​ible
vector ||v|| v2)​2...) genera​l/a​lge​braic equation Ax+By+Cz =
if A is invert​ible: (AB)-1 = B-1A-1
||v|| = 0 iff v =0 ||kv|| = |k| ||v|| D
(An) -1 = A-n = (A-1) n (AT) -1 = (A-1) T
unit vector u in same u = (1/ ||v||) v two planes are parallel if n1 = kn2,
orthogonal if n1·n2 = 0 (kA)-1 1/k(A ​-1), k≠0
direct as v
e1 = (1,0...) ... en = standard unit
Elementary Matrix and Unifying Theorem
(0,...1) in Rn vector
elementary matrices are invertible
d(u,v) = sqrt((​u1-​v1)2 + (u2-v2)2 ... (un-
A-1 = Ek Ek-1 ... E2 E1
vn​)2) = ||u-v||
[ A | ] -> [ | A -1 ]
d(u,v) = 0 iff u = v
(how to find inverse of A)
Ax = b; x = A -1b

By fionaw Published 16th July, 2020. Sponsored by Readable.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Measure your website readability!
Page 1 of 4. https://readable.com
Linear Algebra - MATH 232 Cheat Sheet
by fionaw via cheatography.com/124375/cs/23750/

Elementary Matrix and Unifying Theorem Examples of Subspaces Linear Indepe​ndent (cont)
(cont)
IF: w1, w2 are then w1+w2 are - list the vectors as the columns of a matrix,
- A -> RREF = within S within S row reduce it, if many solution, then it is
- A can be express as a product of E and kw1 is within S dependent
- A is invertible - the zero vector 0 it self is a subspace - after RREF, the columns with leading 1's
- Ax = 0 has only the trivial solution are a maxmially linearly indepe​ndent subset
- Rn is a subspace of all vectors
- Ax = b is consistent for every vector b in according to Pivot Theorem
Rn - Lines and planes through the origin are
subspaces
- Ax = b has eactly 1 solution for every b in Diagonal, Triang​ular, Symmetric
Rn - The set of all vectors b such that Ax = b is Matrices
- colum and rowvectors of A are linealy consis​tent, is a subspace
Diagonal all zeros along the
indepe​ndent - If {v1, v2, ...vk} is any set of vectors in R n,
Matrices diagonal
- det(A) ≠ 0 then the set W of all linear combin​ations of
Lower zeros above diagonal
- λ = 0 is not an eigenvalue of A these vector is a subspace
Triangular
- TA is one to one and onto
W = {c1v1 + c2v2 + ... ckvk}; c are within
If not, then all no. Upper zeros below the
real numbers
Triangular diagonal
Consis​tency Symmetric if: AT = A
Span
EAx = Eb -> Rx = b' , where b' = Eb
- the span of a set of vectors { v1, v2, ... vk} Skew-S​ymm​‐ AT = -A
(Ax=b) [ A | b ] -> [ EA | Eb ] (Rx = b') is the set of all linear combin​ations of these etric if:
(but treat b as unknown: b1, b2...) vectors
Determ​inants
For it to be consis​tent, if R has zero span { v1, v2, ... vk} = { v11t, t2v2, ... , tkvk}
rows at the bottom, b' that row must det(A) = a1jC1j + expansion along
If S = { v1, v2, ... vk}, then W = span(S) is a
equal to zero a2jC2j ... + anjCnj jth column
subspace
det(A) = ai1Ci1 + expansion along
Ax = b is consistent if and only if b is a
Homoge​neous Systems ai2Ci2 ... + ainCin the ith row
linear combin​ation of col(A)
Linear Combin​ation of the vectors: Cij = (-1)i+j Mij
v = c1v1 + c2v2 ... + cnvn Linear Indepe​ndent Mij = deleted ith row and jth column matrix
(use matrix to find c)
- if unique solution for a set of vectors, then - pick the one with most zeros to calculate
Ax = 0 Homoge​neous
it is linearly indepe​ndent easier
Ax = b Non-ho​mog​‐
c1v1 + c2v2 ... + cnvn = 0; all the c = 0 det(AT) = det(A) det(A-1) =
enous
- for dependent, not all the c = 0 1/det(A)
x = x0 + t1v1 + t2v2 ... + Homoge​neous
Dependent if: det(AB) = det(A)​det(B) det(kA) =
tkvk
- a linear combin​ation of the other vectors kndet(A)
x = t1v1 + t2v2 ... + tkvk Non-ho​mog​‐
- a scalar multiple of the other - A is invertible iff det(A) not equal 0
eneous
- a set of more than n vectors in Rn
- det of triangular or diagonal matrix is the
xp is any solution of NH x = xp + xh
Indepe​ndent if: product of the diagonal entries
system
- the span of these two vectors form a plane det(A) for 2x2 matrix ad - bc
and xh is a solution of H
system

By fionaw Published 16th July, 2020. Sponsored by Readable.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Measure your website readability!
Page 2 of 4. https://readable.com
Linear Algebra - MATH 232 Cheat Sheet
by fionaw via cheatography.com/124375/cs/23750/

Adjoint and Cramer's Rule Complex Number Linear Transf​orm​ation

adj(A) = CT CT = matrix confactor complex number a + ib f: Rn -> Rm, n = domain, m = co-dom​ain


of A (a + ib) + (c + id) = (a + c) + i(b + d) f(x1, x2, ...xn) = (y1, ...ym)
A-1 = (1/det(A)) adj(A)A = det(A) I (a + ib) - (c + id) = (a - c) + i(b - d) T: Rn -> Rm is a linear transf​ormatin if
adj(A) 1. T(cu) = cT(u)
(a + ib) (c + id) = (ac + bd) + i(ad + bc)
x1 = det(A1) / x2 = det(A2) / det(A) 2. T(u +v) = T(u + T(v)
(a + bx) (c + dx) = (ac + bdx2) + x(ad + bc)
det(A) for any linear transf​orm​ation, T(0) = 0
i2 = -1
xn = det(An) / det(A) not equal 0 Rθ = [T(e1) T(e2)] = [cosθ matrix for
z = a + ib z bar = a - ib
det(A) −sinθ] rotation
the length​(ma​gni​tude) of |z| = sqrt(z x z ​[sinθ
An is the matrix when the nth column is
vector z bar) cosθ]
replaced by b
= sqrt​(a 2 + b2)
reflection across y-axis: T(x, y) = (-x, y)
Hyperp​lane, Area/V​olume z-1 = 1/z = z bar / |z|2 reflection across x-axis: T(x, y) = (y, -x)
a hyperplane in a1x1 + a2x2 ... + anxn = z1 / z2 = z1z2-1 reflection across diagonal y = x, T(x, y) = (y,
Rn b z = |z| (cos(θ) + i (sin(θ)) polar form (r = x)
- can also written as ax = b |z|) orthogonal projection onto the x-axis: T(x, y)
to find aperp ax = 0, find the span z1z2 = |z1| |z2| (cos(θ1 + θ2) + i (sin(θ1 + = (x, 0)
θ2)) orthogonal projection onto the y-axis: T(x, y)
if A is 2x2 matrix:
- |det(A)| is the area of parall​elogram z1/z2 = |z1| / |z2| (cos(θ1 - θ2) + i (sin(θ1 - = (0, y)
θ2)) u = (1/ ||v||)v; express it vertically as u1 and
if A is 3x3 matrix:
- |det(A)| is the volume of parall​ele​piped zn = rn(cos(n θ) + i sin(n r = |z| u2
θ)) A = [(u1)2 u2u1] projection
- subtract points to get three vectors, then
make it to a matrix to find the area/v​olume ei theta = cos(θ) + i sin(θ) ​[u1u2 (u2)2] matrix

ei pi = -1 ei pi +1 = 0 contra​ction with 0 ≤ k < 1 (shrink), k > 1


Cross Product (stretch)
z1z2 = r1r2 ei (θ1 + θ2) zn = rn ei nθ
u x v = (u2v3 - u3v2, u3v1 - u1v3, u1v2 - [x, y] -> [kx, ky]
z1 /z2 = r1 / r2 e i (θ1 - θ2)
u2v1) compre​ssion in x-dire​ction [x, y] -> [kx, y]

u x v = -v x k(u x v) = (ku) x v = u x (kv) Eigenv​alues and Eigenv​ectors compre​ssion in y-dire​ction [x, y] -> [x, ky]
u shear in x-dire​ction T(x,y) = (x+ky, y);
Ax= λx
uxu=0 parallel vectors has 0 for c.p. [x+ky (1, k), y( 0, 1)]
det(λI - A) = (-1) n det(A - λI)
u (u x v) = 0 v (u x v) = 0 shear in y-dire​ction T(x,y) = (x, y+kx);
pa(λ) = 3x3: det(A - λI); 2x2: det(λI - A)
u x v is perpen​dicular to span {u, v} [x (1, 0), y (k, 1)]
- solve for (λI - A)x = 0 for eigenv​ectors
||u x v|| = ||u|| ||v|| sin(th​eta), where theta is orthogonal projection on the xy-plane: [x, y ,
Work Flow: 0]
the angle between vectors
- form matrix
orthogonal projection on the xz-plane: [x, 0 ,
- compute pa(λ) = det(λI - A)
z]
- find roots of pa(λ) -> eigenv​alues of A
orthogonal projection on the yz-plane: [0, y ,
- plug in roots then solve for the equation
z]
reflection about the xy-plane: [x, y, -z]
reflection about the xz-plane: [x, -y, z]
reflection about the yz-plane: [-x, y, z]

By fionaw Published 16th July, 2020. Sponsored by Readable.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Measure your website readability!
Page 3 of 4. https://readable.com
Linear Algebra - MATH 232 Cheat Sheet
by fionaw via cheatography.com/124375/cs/23750/

Orthogonal Transf​orm​ation Kernel, Range, Compos​ition (cont)

an orthogonal transf​orm​ation is a linear [T(θ1+θ2)] = [Tθ2] ◦ [Tθ1];


transf​orm​ation T; Rn -> Rn that preserves rotate then shear ≠ shear then rotate
lengths; ||T(u)|| = ||u|| linear trans T: Rn->Rm has an inverse iff T
||T(u)|| = ||u|| <=> T(x)·T(y) = x·y for all x,y is one to one, T -1: Rm -> Rn, Tx = y <=> x =
in Rn T -1y
orthogonal matrix is square matrix A such for Rn to Rn, [T-1] = [T]-1; [T]-1◦T = 1n <=>
that AT = A-1 [T-1​][T]= n
1. if A is orthog​onal, then so is AT and A-1 1n is identity transf​orm​ation; n is identity
matrix
2. a product of orthonal matrices is
orthogonal
Basis, Dimension, Rank
3. if A is orthog​onal, then det(A) = 1 or -1
S is a basis for the subspace V of Rn if:
4. if A is orthog​onal, then rows and columns
S is linearly idenpe​ndent and span(S) = V
of A are each orthon​ormal sets of vectors
dim(V) = k, k is the # of vectors
Kernel, Range, Compos​ition row(A) = rows with leading ones after RREF

ker(T) is the set of all vectors x such that col(A) = columns with leading ones from
T(x) = 0, RREF matrix, find the vector, original A
ker(T) = span{(​v)} null(A) = free variable's vectors
the solution space of Ax = 0 is the null null(AT) = after transform, the free variable
space; vector
null(A) = ker(A)
The Rank Theorem: rank(A) = rank(AT) for
range of T, ran(T) is the set of vectors y any matrix have the same dimension
such that
rank(A) = # of free vectors in span
y = T(x) for some x
dim(ro​w(A)) = dim(co​l(A)) = rank(A)
ran(T) = col([T]) = span{ [col1], [col2] ...}; Ax
dim(nu​ll(A)) = nullity(A)
=b
Important Facts:
Orthogonal Compli​ment, DImention
1. T is one to one iff ker(T) = {0}
Theorem
2. Ax = b, if consis​tent, has a unique
solution S⟂ = {v ∈ Rn | v · w = 0 for all w ∈ S}
iff null(A) = {0}; Ax = 0 has only the trivial S⟂ is a subspace of Rn; S⟂ = span(S)⟂ = W⟂
solution iff null(A) = {0}
row(A)⟂ = null(A) null(A)⟂ = row(A)
Important facts 2: ((S ⟂) ⟂ = S iff S is
1.T:Rn -> Rm is onto iff the system Tx = y subspace
has a solution x in Rn for every y in Rm
col(A) ⟂ = null(AT) null(A​T) ⟂ = col(A)
2. Ax = b is consistent for every b in R m(A is
The Dimension rank(A) + nullity(A) =
onto) iff col(A) = Rm
Theorem n
The compos​ition of T2 with T1 is: T2 ◦ T1 A is m x n matrix (k + (n-k) = n)
(T2 ◦ T1)(x) = T2(T1(x)); T2 ◦ T1: Rn -> Rm if W is a subspace dim(W) + dim(W⟂) =
compostion of linear transf​orm​ations corres​‐ of Rn n
ponds to matrix applic​ation: [T2 ◦ T1] = [T1]
[T2]

By fionaw Published 16th July, 2020. Sponsored by Readable.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Measure your website readability!
Page 4 of 4. https://readable.com

You might also like