Beruflich Dokumente
Kultur Dokumente
2003 - Introduction To Continuum Mechanics - Vector and Tensor Calculus - Franz-Joseph Barthold
2003 - Introduction To Continuum Mechanics - Vector and Tensor Calculus - Franz-Joseph Barthold
Braunschweig
Introduction to
Continuum Mechanics
Franz-Joseph Barthold 1
Jrg Stieghan 2
1 Tel.
2 Tel.
Herausgeber
Abstract
Zusammenfassung
Organisation und Verwaltung
Dipl.-Ing. Jrg Stieghan, SFI
CSE Computational Sciences in Engineering
Technische Universitt Braunschweig
Bltenweg 17, 38 106 Braunschweig
Tel. ++49-(0)531-391-2247
Fax ++49-(0)531-391-2242
email j.stieghan@tu-bs.de
c
2000
Alle Rechte, insbesondere das der bersetzung in fremde Sprachen, vorbehalten. Ohne Genehmigung der Autoren ist es nicht gestattet, dieses Heft ganz oder teilweise auf fotomechanischem
Wege (Fotokopie, Mikroskopie) zu vervielfltigen oder in elektronische Medien zu speichern.
Preface
Contents
Contents
VII
List of Figures
IX
List of Tables
XI
Introduction
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3
6
8
10
12
16
18
25
28
32
36
Matrix Calculus
3.1 Denitions . . . . . . . . . . . . . . . . . .
3.2 Some Basic Identities of Matrix Calculus .
3.3 Inverse of a Square Matrix . . . . . . . . .
3.4 Linear Mappings of an Afne Vector Spaces
3.5 Quadratic Forms . . . . . . . . . . . . . .
3.6 Matrix Eigenvalue Problem . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
37
40
42
48
54
62
65
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
75
78
85
96
101
VII
.
.
.
.
Contents
VIII
4.5
4.6
4.7
5
131
133
143
152
Exercises
6.1 Application of Matrix Calculus on Bars and Plane Trusses
6.2 Calculating a Structure with the Eigenvalue Problem . . .
6.3 Fundamentals of Tensors in Index Notation . . . . . . . .
6.4 Various Products of Second Order Tensors . . . . . . . . .
6.5 Deformation Mappings . . . . . . . . . . . . . . . . . . .
6.6 The Moving Trihedron, Derivatives and Space Curves . . .
6.7 Tensors, Stresses and Cylindrical Coordinates . . . . . . .
159
162
174
182
190
194
198
210
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
A Formulary
227
A.1 Formulary Tensor Algebra . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 227
A.2 Formulary Tensor Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233
B Nomenclature
237
References
239
241
257
Index
273
List of Figures
2.1
2.2
2.3
2.4
2.5
Triangle inequality. . . . . . . . . . . . . . . . . . . . . . . .
Hlder sum inequality. . . . . . . . . . . . . . . . . . . . . .
Vector space R2 . . . . . . . . . . . . . . . . . . . . . . . . . .
Afne vector space R 2af f ine . . . . . . . . . . . . . . . . . . .
The scalar product in an 2-dimensional Euclidean vector space.
3.1
3.2
3.3
Matrix multiplication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
Matrix multiplication for a composition of matrices. . . . . . . . . . . . . . . . . 55
Orthogonal transformation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
4.1
4.2
4.3
4.4
4.5
4.6
4.7
4.8
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
81
82
86
96
97
117
119
120
5.1
5.2
5.3
5.4
5.5
5.6
5.7
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
136
137
138
140
141
152
154
6.1
6.2
6.3
6.4
6.5
6.6
6.7
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
162
162
163
164
165
166
167
IX
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
16
21
28
28
30
List of Figures
X
6.8
6.9
6.10
6.11
6.12
6.13
6.14
6.15
6.16
6.17
. 174
.
.
.
.
.
.
.
.
.
175
176
182
184
198
199
204
207
210
List of Tables
2.1
Compatibility of norms. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
XI
XII
List of Tables
Chapter 1
Introduction
Chapter 1. Introduction
Chapter 2
Basics on Linear Algebra
For example about vector spaces H ALMOS [6], and A BRAHAM, M ARSDEN, and R ATIU [1].
And in german DE B OER [3], and S TEIN ET AL . [13].
In german about linear algebra JNICH [8], F ISCHER [4], F ISCHER [9], and B EUTELSPACHER
[2].
Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.1.1
2.1.2
2.1.3
2.2
2.4
2.5
2.2.1
Denition of a Mapping . . . . . . . . . . . . . . . . . . . . . . . . .
2.2.2
2.2.3
Denition of an Operation . . . . . . . . . . . . . . . . . . . . . . . .
2.2.4
Examples of Operations . . . . . . . . . . . . . . . . . . . . . . . . .
Counter-Examples of Operations . . . . . . . . . . . . . . . . . . . . .
2.7
2.8
Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.3.1
Denition of a Field . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.3.2
Examples of Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.3.3
Counter-Examples of Fields . . . . . . . . . . . . . . . . . . . . . . . 11
Linear Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.4.1
2.4.2
2.4.3
2.4.4
2.4.5
Linear Independence . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.4.6
Metric Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.6.7
2.6.8
7
8
2.5.1
2.6
Examples of Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Mappings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2.5
2.3
2.6.6
2.9
2.7.2
2.7.3
2.7.4
2.7.5
Unitary Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
2.8.2
2.8.3
2.9.2
2.9.3
2.9.4
2.9.5
2.9.6
2.9.7
Denition of a Metric . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.5.2
Examples of Metrices . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2.5.3
2.5.4
Normed Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.6.1
Denition of a Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.6.2
2.6.3
2.6.4
2.6.5
Matrix Norms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.1
Sets
2.1. Sets
A B (a A a B) .
The set B is called the superset of A
{. . . | . . .} : The terms on the left-hand side of the vertical bar are the elements of the
given set and the terms on the right-hand side of the bar describe the characteristics of the
elements include in this set.
(2.1.5)
The union C of two sets A and B is the set of all elements, that at least are an element of one of
the sets A and B
C = A B = {c | (c A) (c B)} .
(2.1.6)
C = A B = {c | (c A) (c B)} .
(2.1.7)
= { }.
B A.
(2.1.4)
The intersection C of two sets A and B is the set of all elements common to the sets A and B
(2.1.8)
Example: The set of natural numbers. The set of natural numbers, or just the naturals, N,
sometimes also the whole numbers, is dened by
N = {1, 2, 3, . . .} .
(2.1.9)
M1 = {1, 2, 3, } .
(2.1.1)
Unfortunately, zero "0"is sometimes also included in the list of natural numbers, then the set N
is given by
(2.1.10)
N0 = {0, 1, 2, 3, . . .} .
N = {1, 2, 3, . . .} .
(2.1.2)
Z = {z | (z = 0) (z N) (z N)} .
The set N includes all integers larger or equal to one and it is also called the set of natural
numbers.
the description of the attributes of its elements, e.g.
M2 = {m | (m M1 ) (m M1 )} ,
= {1, 2, 3, 1, 2, 3} .
(2.1.11)
Example: The set of rational numbers. The set of rational numbers Q is described by
o
nz
| (z Z) (n N) .
(2.1.12)
Q=
n
Example: The set of real numbers. The set of real numbers is dened by
(2.1.3)
The set M2 includes all elements m with the attribute, that m is an element of the set M 1 , or
that m is an element of the set M1 . And in this example these elements are just 1, 2, 3 and
1, 2, 3.
1
R = {. . .} .
Example: The set of complex numbers. The set of complex numbers is given by
C = + i | (, R) i = 1 .
2
The expression "if and only if" is often abbreviated with "iff".
(2.1.13)
(2.1.14)
2.2
Mappings
The mappings idV : V V and idW : W W are the identity mappings in V, and W, i.e.
idV (x) = x x V
(2.2.3)
(2.2.4)
2.2. Mappings
idW (y) = y
y W .
(2.2.10)
Furthermore f must be surjective, in order to expand the existence of this mapping f 1 from
f (V) W to the whole set W. Then f : V W is bijective, if and only if the mapping
g : W V with g f = idV and f g = idW exists. In this case is g = f 1 the inverse.
x=y
, and
(2.2.5)
(2.2.6)
(2.2.7)
(2.2.8)
; f f 1 = idW .
Counter-Example: The subtraction of natural numbers. The subtraction over the natural
numbers N is not an operation, because there exist numbers a N and b N with a difference
(a b) 6 N. E.g. the difference 3 7 = 4 6 N.
Counter-Example: The scalar multiplication of a n-tuple. The scalar multiplication of a ntuple of real numbers in Rn with a scalar quantity a R is not an operation, because it does not
map two elements of Rn onto another element of the same space, but one element of R and one
element of Rn .
Counter-Example: The scalar product of two n-tuples. The scalar product of two n-tuples
in Rn is not an operation, because it does not map an element of Rn onto an element Rn , but
onto an element of R.
(2.2.9)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
10
2.3
Fields
2.3. Fields
a, b, c F .
a + b = b + a a, b F .
(F2)
a + 0 = a = 0 + a a F .
(F3)
3 . Axiom of Fields. There exists a unique scalar 0 F, called zero or the identity element with
respect to3 the addition of the eld F, such that the additive identity is given by
4 . Axiom of Fields. To every scalar a F there corresponds a unique scalar a, called the
inverse w.r.t. the addition or additive inverse, such that
a + (a) = 0
a F .
Example: The rational numbers. The set Q of the rational numbers together with the operations addition "+"and multiplication ""describe a eld.
Example: The real numbers. The set R of the real numbers together with the operations addition "+"and multiplication ""describe a eld.
Example: The complex numbers. The set C of the complex numbers together with the operations addition "+"and multiplication ""describe a eld.
(F1)
Counter-Example: The natural numbers. The set N of the natural numbers together with the
operations addition "+"and multiplication ""do not describe a eld! One reason for this is that
there exists no inverse w.r.t. the addition in N.
Counter-Example: The integers. The set Z of the integers together with the operations addition "+"and multiplication ""do not describe a eld! For example there exists no inverse w.r.t.
the multiplication in Z, except for the elements 1 and 1.
(F4)
To every pair, a and b, of scalars there corresponds a scalar ab, called the product of a and b, in
such way that:
5 . Axiom of Fields. The multiplication is associative ,
a (bc) = (ab) c a, b, c F .
(F5)
(F6)
7 . Axiom of Fields. There exists a unique non-zero scalar 1 F, called one or the identity
element w.r.t. the multiplication of the eld F, such that the scalar multiplication identity is given
by
a1 = a = 1a a F .
(F7)
1
a a1 = 1 = a
a F .
(F8)
a
9 . Axiom of Fields. The muliplication is distributive w.r.t. the addition, such that the distributive law is given by
(a + b) c = ac + bc a, b, c F .
(F9)
3
11
12
2.4
Linear Spaces
13
7 . Axiom of Linear Spaces. The scalar muliplication is distributive w.r.t. the vector addition,
such that the distributive law is given by
(x + y) = x + y
F ;
x, y V .
(S7)
8 . Axiom of Linear Spaces. The muliplication by a vector is distributive w.r.t. the scalar
addition, such that the distributive law is given by
( + ) x = x + x , F
; x V .
(S8)
(2.4.2)
and satises the following axioms. The elements x, y etc. of the V are called vectors. To every
pair, x and y of vectors in the space V there corresponds a vector x + y, called the sum of x and
y, in such a way that:
1 . Axiom of Linear Spaces. The addition is associative ,
x + (y + z) = (x + y) + z x, y, z V .
0 x = 0 x V ; 0 F,
(1) x = x x V ; 1 F,
0 = 0 F,
(2.4.3)
(2.4.4)
(2.4.5)
and if
(S1)
x = 0
, then
=0
, or
x = 0.
(2.4.6)
(S2)
3 . Axiom of Linear Spaces. There exists a unique vector 0 V, called zero vector or the
origin of the space V, such that
x + 0 = x = 0 + x x V .
(S3)
x V .
(S4)
To every pair, and x, where is a scalar quantity and x a vector in V, there corresponds a
vector x, called the product of and x, in such way that:
5 . Axiom of Linear Spaces. The multiplication by scalar quantities is associative
(x) = () x , F ; x V .
(S5)
6 . Axiom of Linear Spaces. There exists a unique non-zero scalar 1 F, called identity or the
identity element w.r.t. the scalar multiplication on the space V, such that the scalar multplicative
identity is given by
x1 = x = 1x x V .
(S6)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
2.4.1.0.1
Remarks:
Starting with the usual 3-dimensional vector space these axioms describe a generalized
denition of a vector space as a set of arbitrary elements x V. The classic example is
the usual 3-dimensional Euclidean vector space E3 with the vectors x, y.
The denition says nothing about the character of the elements x V of the vector space.
The denition implies only the existence of an addition of two elements of the V and the
existence of a scalar multiplication, which both do not lead to results out of the vector
space V and that the axioms of vector space (S1)-(S8) hold.
The denition only implies that the vector space V is a non empty set, but nothing about
"how large"it is.
F = R, i.e. only vector spaces over the eld of real numbers R are examined, no look at
vector spaces over the eld of complex numbers C.
The dimension dim V of the vector space V should be nite, i.e. dim V = n for an arbitrary
n N, the set of natural number.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
14
(2.4.7)
(2.4.11)
Furthermore let M = {x1 , x2 , . . . , xm } be a set of vectors. Than the set of all linear combinations
of the vectors x1 , x2 , . . . , xm is called the span span (M) of the subspace M and is dened by
span (M) = a1 x1 + a2 x2 + . . . + am xm | a1 , a2 , . . . am F .
(2.4.12)
x1 , x2 , . . . , xn R.
Let V be a linear space over the eld F with the vectors x 1 , x2 , . . . , xm V. Every vector v V
could be represented by a so called linear combination of the x 1 , x2 , . . . , xm and some scalar
quantities a1 , a2 , . . . , am F
v = a 1 x1 + a2 x2 + . . . + a m xm .
Example: The space of n n-matrices. The space of square matrices Rnn over the eld R
with the usual matrix addition and the usual multiplication of a matrix with a scalar quantity is a
linear space over the eld R, denoted by
a11 a12
a1n
a21 a22
a2n
A = ..
; aij R , 1 i m, 1 j n , and i, j N.
..
.
.
. . ..
.
.
am1 am2 amn
(2.4.8)
Example: The eld. Every eld F with the deniton of an addition of scalar quantities in the
eld and a multiplication of the scalar quantities, i.e. a scalar product, in the eld is a linear
space over the eld itself.
Example: The space of continous functions. The space of continuous functions C (a, b) is
given by the open intervall (a, b) or the closed intervall [a, b] and the complex-valued function
f (x) dened in this intervall,
C (a, b) = {f (x) | f
Let V be a linear space over the eld F. A subset W V is called a linear subspace or a linear
manifold of V, if the set is not empty, W 6= , and the linear combination is again a vector of the
linear subspace,
ax + by W x, y W ; a, b F .
(2.4.10)
x = [x1 , . . . , xn ] ,
15
[a, b]} ,
(2.4.9)
Let V be a linear space over the eld F. The vectors x 1 , x2 , . . . , xn V are called linearly
independent, if and only if
n
X
ai xi = 0 = a1 = a2 = . . . = an = 0.
n
X
v i ei ,
i=1
(f + g) = f (x) + g (x) ,
(f ) = f (x) .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(2.4.13)
i=1
(2.4.15)
16
2.5
Metric Spaces
17
V V F
(x, y) 7 (x, y)
(2.5.1)
x, y V .
(x, y) 0
(M1)
Example: The distance in the Euclidean space. For two vectors x = (x 1 , x2 )T and y =
(y1 , y2 )T in the 2-dimensional Euclidean space E2 the distance between this two vectors, given
by
q
(x, y) =
(x1 y1 )2 + (x2 y2 )2
is a metric.
Example: Discrete metric. The mapping, called the discrete metric,
(
0, if x = y
(x, y) =
,
1, else
(x, y) = 0 x = y
(M2)
(M3)
x, y, z V .
(M4)
Example: The eld. The eld of the complex numbers C is a metric space.
Example: The vector space. The vector space Rn is a metric space, too.
y
(x, y)
(y, z)
x
(x, z)
z
Figure 2.1: Triangle inequality.
(2.5.3)
(2.5.2)
(2.5.4)
18
2.6
Normed Spaces
19
kx + yk kxk + kyk .
(2.6.1)
x V .
Example: The normed vector space. For the linear vector space Rn , with the zero vector 0,
there exists a large variety of norms, e.g. the l-innity-norm, maximum-norm,
(N1)
kxk = max |xi |
x V .
(N2)
F ;
x V .
x, y V .
kxk1 =
(N3)
(N4)
kxk =
n
X
|xi | ,
(2.6.9)
|x| d,
(2.6.10)
i=1
(2.6.11)
i=1
the L2-norm,
vZ
u
u
kxk = t |x|2 d,
The norm of a vector x is written like kxk and is called the vector norm. For a vector norm the
following conditions hold, see also (N1)-(N4),
x 6= 0,
(2.6.4)
, and R ,
(2.6.5)
(2.6.12)
(2.6.8)
kxk = kxk ,
kxk kyk kx yk .
1 i n,
the L1-norm,
, with
the l1-norm,
(2.6.7)
i=1
The norm satises the following relations for all vectors x, y, z V and every F:
1 . Axiom of Norms. The norm is positive,
kxk 0
(2.6.6)
kxk =
n
X
i=1
|xi |
! p1
, with
1 p < .
(2.6.13)
20
21
y 6
i = 1, . . . , n,
i=1
n
X
i=1
|xi |p
! p1
p
nz.
(2.6.14)
Figure 2.2: Hlder sum inequality.
Example: Simple Example with Numbers. The varoius norms of a vector x differ in most
general cases. For example with the vector xT = [1, 3, 4]:
kxk1 = 8,
kxk2 = 26 5, 1,
kxk = 4.
(2.6.15)
(2.6.16)
For the real or complex quantities xj , and yj , which are not all equal to zero, the , and could
be described by
|yj |
|xj |
(2.6.17)
=
p1 , and = P
1 .
P
p
q q
j |xj |
j |yj |
Inserting the relations of equations (2.6.17) in (2.6.16), and summing the terms with the index j,
implies
P
P
P
p
q
j |xj | |yj |
j |xj |
j |yj |
= 1.
+
(2.6.18)
P
p1 P
1q
P
P
p
p
q
p
|xj |
q
|yj |q
j
j
|x
|
|y
|
j
j
j
j
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
X
j
In the rst quadrant of a coordinate system the graph y = x p1 and the straight lines x = , and
y = with > 0, and > 0 are displayed. The area enclosed by this two straight lines, the
curve and the axis of the coordinate system is at least the area of the rectangle given by ,
p q
+ .
p
q
|xj yj |
X
j
|xj |
! p1
X
j
|yj |
! 1q
(2.6.19)
For the special case with p = q = 2 the Hlder sum inequality, see equation (2.6.19) transforms
into the Cauchys inequality,
Let p and q be two scalar quantities, and the relationship between them is dened by
p > 1, q > 1.
|xj yj |
X
j
|xj |
! 12
X
j
|yj |
! 12
(2.6.20)
(2.6.21)
(2.6.22)
(2.6.23)
(2.6.24)
In addition for the matrix norms and in opposite to vector norms the last axiom hold. If this
condition holds, then the norm is called to be multiplicative. Some usual norms, which satisfy
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
22
the conditions (2.6.21)-(2.6.22) are given below. With n being the number of rows of the matrix
A, the absolute norm is given by
kAkM = M (A) = n max |aik | .
n
X
k=1
|aik | .
(2.6.26)
n
X
i=1
|aik | .
kAkN = N (A) =
The spectral norm is given by
kAkH = H (A) =
tr A A .
T
largest eigenvalue of AT A .
Description
absolute norm
maximum absolute row sum norm
kxk =
kAkM = M (A)
kAkC = C (A) = sup (A)
absolute norm
maximum absolute column sum norm
kxk =
qP
kAkM = M (A)
kAkN = N (A)
kAkH = H (A) = sup (A)
absolute norm
Euclidean norm
spectral norm
|xi |
|xi |2
(2.6.28)
Denition 2.2. The supremum sup (x) of a matrix A associated to the vector norm kxk is dened
by the scalar quantity , in such a way that,
(2.6.29)
kAxk kxk ,
(2.6.32)
(2.6.33)
Denition 2.1. A matrix norm kAk is called to be compatible to an unique vector norm kxk, iff
for all matrices A and all vectors x the following inequality holds,
kA xk kAk kxk .
(2.6.27)
Vector norms
kxk = max |xi |
(2.6.25)
23
(2.6.30)
or
sup (A) = max
kA xk
.
kxk
(2.6.34)
i ai = 0.
(2.6.35)
i=1
This equations implies immediately, that the matrix norm is an estimation of the eigenvalues.
Then with this condition a compatible matrix norm associated to a vector norm is most valuable,
if in the inequality kA xk kAk kxk, see also (2.6.31), both sides are equal. In this case there
can not exist a value of the left-hand side, which is less than the value of the right-hand side.
This upper limit is called the supremum and is written like sup (A).
In every other case the vectors are called to be linearly dependent. For example three linearly
independent vectors are given by
1
0
0
1
2
3
0 + 1 + 0 6= 0 , with i =
6 0.
(2.6.36)
0
0
1
24
The n linearly independent vectors ai with i = 1, . . . , n span a n-dimensional vector space. This
set of n linearly independent vectors could be used as a basis of this vector space, in order to
describe another vector an+1 in this space,
an+1 =
n
X
k=1
k ak
, and
an+1 Rn .
(2.6.37)
2.7
25
x, y V .
(I1)
(I2)
(I3)
(I4)
(2.7.2)
Theorem 2.1. The inner product induces a norm and with this a metric, too. The scalar product
1
kxk = hx, xi 2 denes a scalar-valued function, which satises the axioms of a norm!
(2.7.3)
It is important to notice, that the scalar product and the scalar multiplication are complete different mappings!
26
27
A vector space V with a scalar product h , i is called an inner product space or Euclidean vector
space 5 . The axioms (N1), (N2), and (N3) hold, too, only the axiom (N4) has to be proved. The
axiom (N4) also called the Schwarz inequality is given by
A vector space V over the eld of real numbers R, with a scalar product h , i is called an inner
product space, and sometimes its complex analogue is called an unitary space over the eld of
complex numbers C.
(2.7.4)
(2.7.5)
hu, ui = uT Au,
(2.7.6)
(2.7.7)
28
2.8
In matrix calculus an n-tuple a Rn over the eld of real numbers R is studied, i.e.
, and i = 1, . . . , n.
V R2
~ P~ )
(R
P Q + QR = P R,
(2.8.5)
P Q = P Q
, and Q W.
(2.8.6)
For all P , Q and R W Rnafne the axioms of a linear space (S1)-(S4) for the addition hold
~ P~ )
(Q
~a = P Q
~b =
QR
P : W V , with
~c = P R
(2.8.4)
and
P : Q P Q,
(2.8.1)
One of this n-tuple, represented by a column matrix, or also called a column vector or just vector,
could describe an afne vector, if an point of origin in a geometric sense and a displacement of
origin are established. A set W is called an afne vector space over the vector space V R n , if
6
29
assigns to every pair of points P and Q W Rnafne a vector P Q V. And the mapping also
satises the following conditions:
ai R
a + b = c ai + bi = ci
, with
i = 1, . . . , n,
(2.8.7)
P~
a = a ai = ai .
(2.8.8)
W R2afne
An Euclidean vector space En is an unitary vector space or an inner prodcut space. In addition
to the normed spaces there is an inner product dened in an Euclidean vector space. The inner
product assigns to every pair of vectors u and v a scalar quantity ,
~b
~c
Q
P
hu, vi u v = v u =
~a
, and
R.
(2.8.9)
a mapping given by
u, v En
For example in the 2-dimensional Euclidean vector space the angle between the vectors u and
v is given by
uv
u v = |u| |v| cos , and cos =
.
(2.8.10)
|u| |v|
W W V ,
Rnafne Rn ,
, with
(2.8.2)
(2.8.3)
Two normed space V and W over the same eld are isomorphic, if and only if there exists
a linear mapping f from V to W, such that the following inequality holds for two constants
m and M in every point x W,
m kxk kf (x)k M kxk .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(2.8.11)
30
31
The set of all linear comnbinations of vectors v1 , v2 , . . . , vn span a subspace. The dimension of this subspace is equal to the number of vectors n, which span the largest linearly
independent space. The dimension of this subspace is at most n.
Every n + 1 vectors of the Euclidean vector space v En with the dimension n must be
linearly dependent, i.e. the vector v = vn+1 could be described by a linear combination of
the vectors v1 , v2 , . . . , vn ,
u
O
Figure 2.5: The scalar product in an 2-dimensional Euclidean vector space.
Every two real n-dimensional normed spaces are isomorphic. For example two subspaces
of the vector space Rn .
Bellow in most cases the Euclidean norm with p = 2 is used to describe the relationships between
the elements of the afne (normed) vector space x R nafne and the elements of the Euclidean
vector space v En . With this condition the relations between a norm, like in section (2.6) and
an inner product is given by
kxk2 = x x,
(2.8.12)
and
q
kxk = kxk2 = x2i .
v + a1 v1 + a2 v2 + . . . + an vn = 0,
1
v = a1 v 1 + a 2 v 2 + . . . + a n v n .
1
z i = ai v i
, with
In this case it is possible to dene a bijective mapping between the n-dimensional afne vector
space and the Euclidean vector space. This bijectivity is called the topology a homeomorphism,
and the spaces are called to be homeomorphic. If two spaces are homeomorphic, then in both
spaces the same axioms hold.
n
X
i=1
v i gi
, with
ai
vi = .
(2.8.18)
The v i gi are called to be the components and the v i are called to be the coordinates of the
vector v w.r.t. the basis gi . Sometimes the scalar quantities v i are called the components
of the vector v w.r.t. to the basis gi , too.
The conditions for the linear dependence and the linear independence of vectors v i in the ndimensional Euclidean vector space En are given below. Furthermore a a vector basis of the
Euclidean vector space En is introduced, and the representation of an arbitrary vector with this
basis is described.
The set of vectors v1 , v2 , . . . , vn is linearly dependent, if there exists a number of scalar
quantities a1 , a2 , . . . , an , not all equal to zero, such that the following condition holds,
(2.8.14)
(2.8.17)
a1 v1 + a2 v2 + . . . + an vn = 0.
(2.8.16)
are called the components of the vector v in the Euclidean vector space E n .
v = v 1 g1 + v 2 g2 + . . . + v n gn =
(2.8.13)
i = 1, . . . , n,
(2.8.15)
32
2.9
Let V and W be two vector spaces over the eld F. A mapping f : V W from elements of
the vector space V to the elements of the vector space W is linear and called a linear mapping,
if for all x, y V and for all R the following axioms hold:
1 . Axiom of Linear Mappings (Additive w.r.t. the vector addition). The mapping f is
additive w.r.t. the vector addition,
x, y V .
(L1)
2 . Axiom of Linear Mappings (Homogeneity of linear mappings). The mapping f is homogeneous w.r.t. scalar multiplication,
f (x) = f (x)
2.9.1.0.2
F ;
x V .
(L2)
Remarks:
The linearity of the mapping f : V W results of being additive (L1), and homogeneous
(L2).
Because the action of the mapping f is only dened on elements of the vector space V, it
is necessary that, the sum vector x + y V (for every x, y V) and the scalar multiplied
vector x V (for every f R) are elements of the vector space V, too. And with this
postulation the set V must be a vector space!
x V ,
(L3)
for all linear mappings f1 , f2 from V to W. The sum f1 + f2 is linear, because both mappings f1
and f2 are linear, i.e. (f1 + f2 ) is a linear mapping, too.
4 . Axiom of Linear Mappings (Denition of the scalar multiplication of linear mappings).
Furthermore a product of a scalar quantity inR and a linear mapping f : V W is dened
by
(f ) (x) := f (x) R ; x V .
(L4)
If the mapping f is linear, then results immediatly, that the mapping (f ) is linear, too.
5 . Axiom of Linear Mappings (Satisfaction of the axioms of a linear vector space). The
denitions (L3) and (L4) satisfy all linear vector space axioms given by (S1)-(S8). This is easy
to prove by computing the equations (S1)-(S8). If V and W are two vector spaces over the eld
F, then the set L of all linear mappings f : V W from V to W,
L (V, W)
(L5)
The identity element w.r.t the addition of a vector space L (V, W) is the null mapping 0, which
sends every element from V to the zero vector 0 W.
With the same arguments for the ranges f (x), f (y), and f (x + y), also for the ranges
f (x), and f (x) in W the set W must be a vector space!
A linear mapping f : V W is also called a linear transformation, a linear operator or a
homomorphism.
33
f (x + y) = f (x) + f (y)
34
35
Till now only an addition of linear mappings and a multiplication with a scalar quantity are
dened. The next step is to dene a "multiplication"of two linear mappings, this combination of
two functions to form a new single function is called a composition. Let f 1 : V W be a linear
mapping and furthermore let f2 : X Y be linear, too. If the image set W of the linear mapping
f1 also the domain of the linear mapping f2 , i.e. W = X, then the composition f1 f2 : V Y
is dened by
(2.9.1)
(f1 f2 ) (x) = f1 (f2 (x)) x V .
Let x and y be two arbitrary elements of the linear vector space V given by
x=
n
X
xi ei
y=
i=1
n
X
y i ei .
(2.9.7)
i=1
(2.9.8)
y = L (x) ,
y i ei = kl kl xj ej
= kl kl xj ej
(2.9.9)
Because of the linearity of the mappings f1 and f2 the composition f1 f2 is also linear.
2.9.4.0.3
, and
Remarks:
(2.9.2)
(2.9.3)
(2.9.4)
(2.9.5)
If all sets are equal V = W = X = Y, then this products exist, i.e. all the linear mappings
map the vector space V onto itself
f L (V, V) =: L (V) .
= kl xj kl (ej )
y=.
(2.9.10)
(2.9.6)
x1
x1
..
..
i
x = x ei . x = . ,
xn
xn
In this case with f1 L (V, V), and f2 L (V, V) the composition f1 f2 L (V, V) is a
linear mapping from the vector space V to itself, too.
with
xV
dim V = n
(2.9.11)
x Rn .
36
2.10
hx, yi =
n
X
x i yi .
Chapter 3
(2.10.1)
i=1
The relations between the continuous linear functionals f : R n R and the scalar products h , i
dened in the R n are given by the Riesz representation theorem, i.e.
Theorem 2.2 (Riesz representation theorem). Every continuous linear functional f : R n R
could be represented by
(2.10.2)
f (x) = hx, ui x Rn ,
Matrix Calculus
For example G ILBERT [5], and K RAUS [10]. And in german S TEIN ET AL . [13], and Z URMHL
[14].
37
38
Denitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
3.1.1
Rectangular Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
3.1.2
Square Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
3.1.3
3.1.4
Row Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Diagonal Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
3.1.6
Identity Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
3.1.7
Transpose of a Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . 41
3.1.9
3.2
Column Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
3.1.5
3.1.8
Symmetric Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
Antisymmetric Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . 41
3.5
3.2.2
3.2.3
Matrix Multiplication . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
3.2.4
3.2.5
3.2.6
3.2.7
3.2.8
3.2.9
3.6
3.4.5
3.4.6
Orthogonal Transformation . . . . . . . . . . . . . . . . . . . . . . . . 57
3.4.7
Quadratic Forms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
3.5.1
3.5.2
3.5.3
3.6.2
Rayleigh Quotient . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
3.6.3
3.6.4
Similarity Transformation . . . . . . . . . . . . . . . . . . . . . . . . 69
3.6.5
3.6.6
Cayley-Hamilton Theorem . . . . . . . . . . . . . . . . . . . . . . . . 71
3.6.7
3.4
3.3.2
3.3.3
3.3.4
3.3.5
3.4.2
3.4.3
3.4.4
39
40
3.1
Denitions
3.1. Denitions
41
A11 A12
A21 A22
A = [Aik ] = ..
..
.
.
Am1
A1n
A2n
.. .
...
.
Amn
(3.1.1)
The index i is the row index and k is the column index. This matrix is called a m n-matrix.
The order of a matrix is given by the number of rows and columns.
Sometimes a diagonal matrix is written like this, because there are only elements on the main
diagonal of the matrix
(3.1.6)
D = d D11 Dmm c.
1 0 0
(
0 1 0
1ik = 0
1 = .. .. . . .. =
. .
. .
1ik = 1
0 0 1
, iff
, iff
i 6= k
i=k
(3.1.7)
A matrix is said to be square, if the number of rows equals the number of columns. It is a
n n-matrix
(3.1.2)
A = [Aik ] = ..
..
.. .
...
.
.
.
An1 Ann
The matrix transpose is the matrix obtained by exchanging the columns and rows of the matrix
A = [aik ]
, and
AT = [aki ] .
a1
a2
T
a = .. = a1 a2 am .
.
am
(3.1.9)
(3.1.8)
(3.1.10)
aik = aki .
a = a1 a 2 a n .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
For that reason a antisymmetric matrix must have zeros on its diagonal.
(3.1.4)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.1.11)
42
3.2
(3.2.1)
A(lm)
43
B (mn)
cij
C (ln)
(3.2.2)
(3.2.3)
And there exists an identity element w.r.t. matrix addition 0, called the additive identity, and
dened by A + 0 = A. Furthermore there exists an inverse element w.r.t. matrix addition A,
called the additive inverse, and dened by A + X = 0 X = A.
(3.2.9)
There is an exception, the so called commutative matrices, which are diagonal matrices of the
same order.
A11 A12
A21 A22
A = A = ..
..
.
.
Am1
A1n
A2n
..
...
.
Amn
(3.2.4)
tr A = tr [Aik ](mn) =
n
X
Aii .
(3.2.10)
i=1
tr (A + B) = tr A + tr B.
(3.2.5)
(3.2.6)
=1
It is important to notice the condition, that the number of columns of the rst matrix equals the
number of rows of the second matrix, see index m in equation (3.2.5). Matrix multiplication is
associative
(A B) C = A (B C) ,
(3.2.7)
(3.2.12)
but still the matrix multiplication in general is not commutative, see equation (3.2.9),
A B 6= B A.
(3.2.13)
(3.2.11)
(3.2.8)
tr 1(nn) = n.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.2.14)
44
Every matrix M could be described as a sum of a symmetric part S and an antisymmetric part A
M (nn) = S (nn) + A(nn) .
(3.2.15)
, i.e.
45
C T = [Cki ]
, and
Cki =
m
X
Ak Bi =
=1
m
X
Bi Ak ,
=1
S ik = S ki .
(3.2.16)
C T = (A B)T = B T AT .
, i.e.
Aik = Aki
, and
Aii = 0.
(3.2.17)
The identity matrix is the multiplicative identity w.r.t. the matrix multiplication
0 1 5
A = 1 0 2
5 2 0.
A 1 = 1 A = A.
1
M + MT +
M M T = S + A.
M=
2
2
The transpose of the symmetric and the antisymmetric part of a square matrix are given by,
T
1
1 T
M + MT =
M + M = S, and
2
2
T
1
1 T
AT =
M MT =
M M = A.
2
2
ST =
(3.2.19)
(3.2.20)
(A B C) = C T B T C T
, etc.
The proof starts with the l n-matrix C, which is given by the two matrices A and B
C (ln) = A(lm) B (mn)
Cik =
m
X
(3.2.24)
Ai Bk .
=1
(3.2.21)
(3.2.22)
(3.2.23)
D11 0
0 D22
D = [Dik ] = ..
..
.
.
0
0
...
0
0
..
.
Dnn
(3.2.25)
(nn)
Because the matrix multiplication is non-commutative, there exists two possibilities two compute
the product of two matrices. The rst possibility is the multiplication with the diagonal matrix
from the left-hand side, this is called the pre-multiplication
D11 a1
a1
D22 a
a
2
2
D A = .. ; A = .. .
(3.2.26)
.
.
Dnn an
an
Each row of the matrix A, described by a so called row vector ai or a row matrix
(3.2.27)
is multiplied with the matching diagonal element Dii . The result is the matrix D A in equation
(3.2.26). The second possibility is the multiplication with the diagonal matrix from the righthand side, this is called the post-multiplication
46
Each column of the matrix A, described by a so called column vector ai or a column matrix
ai1
ai2
(3.2.29)
ai = .. ,
.
ain
is multiplied with the matching diagonal element Dii . The result is the matrix A D in equation
(3.2.28).
47
Vii =
1X
1
Sii = tr S
n i=1
n
, or
V =
1
tr S .
n
The deviator part is the difference between the matrix S and the volumetric part
Rii = Sii Vii
, R=S
1
tr S ,
n
Rik = Sik
, i 6= k
,and R = RT .
(3.2.34)
V = [V ik ] =
V
V
...
V
The matrix T is same as its inverse T = T 1 . And with another matrix T the i-th and the j-th
row are exchanged, too.
j
i
.
.
..
1
..
i 0 1
1
.. . .
..
.
, T = T T
.
.
T =
..
. . . ..
.
.
j 1 0
..
..
.
.
1
Furthermore the old j-th row is multplied by 1. Finally post-multiplication with such a matrix
T exchanges the columns i and j of a matrix.
(3.2.33)
the non-diagonal elements of the deviator are the elements of the former matrix S
Exchanging the i-th and the j-th row of the matrix A is realized by the pre-multiplication with
the matrix T
(nn)
T (nn) A(nn) = A
(3.2.30)
j
i
..
..
a1
1
a11 a12 a1n
.
.
ai
aj
i 0 1
..
..
..
..
.
.
. 1
.
=
...
...
..
..
.
1 .
j 1 0
aj
ai
an
an
..
..
.
. 1
(3.2.32)
(3.2.31)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.2.35)
48
3.3
4. By exchanging two rows (or columns) the sign of the determinant changes.
3.3.2.0.8
; A = [Aik ] .
(3.3.1)
3.3.2.0.9
A1 := X = [Xik ] .
A1 A x = A1 y x = A1 y
, and A1 A = 1.
(3.3.3)
(3.3.5)
(3.3.6)
The solution of the linear equation system could only exist, if and only if the inverse A1 exists.
The inverse A1 of a square matrix A exists, if the matrix is nonsingular (invertible),
i.e. det A 6= 0; or the difference between the rank and the number of columns resp.
rows d = n r of the matrix A must be equal to zero, i.e. the rank r of the matrix
A(nn) must be equal to the number n of columns or rows (r = n). The rank of a
rectangular matrix A(nn) is dened by the largest number of linearly independent
rows (number of rows m) or columns (number of columns n). The smaller value of
m and n is the characteristic value of the rank.
(3.3.8)
Finally the inverse of a matrix is dened by the following relations between a matrix A and its
inverse A1
1 1
= A,
(3.3.4)
A
A1 A = A A1 ,
[Aik ] [Xki ] = 1.
(3.3.7)
(3.3.2)
49
A1k
A2k
A(nn) = a1 a2 ak an
; ak = .. .
(3.3.9)
.
Ank
The ak span a n-dimensional vector space. Than every other vector, the (n + 1)-th vector an+1 =
r Rn , could be described by an unique linear combination of the former vectors ak , i.e. the
vector r Rn is linearly dependent of the n vectors ak Rn . For that reason the linear equation
system
(3.3.10)
A(nn) x(n1) = r(n1) ; r 6= 0 ; r Rn ; x Rn
has an unique solution
A1 := X
, A X = 1.
(3.3.11)
To compute the inverse X from the equation A X = 1 it is necessary to solve n-times the linear
equation system with the unit vector 1j (j = 1, . . . , n) on the right-hand side. Then the j-th
equation system is given by
X1j
0
X2j 0
.
.
.
.
1 2
.X
.
a a ak an kj = ,
1
. .
.. ..
Xnj
0
A X j = 1j ,
X = A1 = X 1 X 2 X j X n
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.3.12)
(3.3.13)
50
and the identity matrix also represented by its column vectors
1 = 1 1 1 2 1 j 1n ,
0
1
.. . .
.
.
0 0
1
0
1 = ..
.
0
0
.. .
.
(3.3.14)
This result is the same like the result by using the determinant expansion by minors. In this
example the determinant is expanded about its rst row. In general the determinant is expanded
about the i-th row like this
n
n
X
X
(3.3.15)
The solutions represented by the vectors X j could be computed with the determinants.
The indices j, k, l, . . . , n are rearranged in all permutations of the numbers 1, 2, . . . , n and I is the
total number of inversions. The determinant det A is established as the sum of all (n!) elements.
In every case there exists the same number of positive and negative terms. For example the
determinant det A of an 3 3-matrix is computed
, or
231
, or
3 1 2,
(3.3.18)
, or
1 3 2.
(3.3.19)
, or
213
1
A= 2
1
4 0
1 1
0 2,
1 1
0 2
2
+ 4 (1)1+2 det
1
2
+ 0 (1)1+3 det
1
1
2
1
,
0
j=1
Aij is a matrix, created by eliminating the i-th row and the j-th column of A. The factor Aij
51
(3.3.22)
The second rule about determinants implies that exchanging the expanded row i by the row k
leads to an linearly dependent matrix,
(3.3.20)
n
X
j=1
i 6= k,
(3.3.23)
52
or
53
Aij Akj = 0 = 0 A if
i 6= k.
(A B)1 (A B) = 1,
(3.3.24)
using equation (3.3.31) this implies
ik =
1,iff
0,iff
i=k
i 6= k
B 1 A1 A B = 1,
1
0
[ik ] = ..
.
0
0
1
.. . .
.
.
0
0
0
.. = 1.
.
and nally
(3.3.25)
B 1 1 B = 1.
3.3.5.0.11
(3.3.26)
(A B C)1 = C 1 B 1 A1 .
j=1
Aij Xjk = ik
A X = 1,
3.3.5.0.12
(3.3.27)
(3.3.33)
j=1
T
T
A A1 = 1 = A A1 = A1 AT ,
Akj
det A
[Xjk ] = A1 .
(3.3.28)
If the matrix is symmetric, i.e. A = AT , the equations (3.3.26) and (3.3.27) imply
Xjk =
Ajk
det A
[Xjk ] = A1 ,
(3.3.29)
3.3.5.0.13
(3.3.30)
3.3.5.0.14
and nally
T
A1 = A1 .
1. The inverse of a matrix prodcut is given by
(A B)
AT
1 1 T
AT = 1 AT
= A
.
A = AT A1 = A1 .
(3.3.34)
(3.3.32)
=B A .
D1
(3.3.31)
n
Y
Dii ,
(3.3.35)
i=1
1
=
.
Dii
(3.3.36)
54
3.4
55
C
i A
Cik
x Rm
with
, y Rl
, and
A Rlm ,
(3.4.1)
yi =
n
X
Aij xj
j=1
...
..
y = A x + . . . + A x + . . . + A x
l
l1 1
lj j
(3.4.2)
lm m
This linear function describes a mapping of the m-tuple (vector) x onto the l-tuple (vector) y
with a matrix A. Furthermore the vector x Rm is described by a linear mapping with a matrix
B and a vector z Rn
x=Bz
, with
xR
,
n
X
zR
, and B R
mn
Bik zk .
(3.4.3)
k=1
k=1
j=1
x=T x
, with x, x
Rn , T Rnn , and
y = T y , with y, y Rn , T Rnn , and
(3.4.5)
(3.4.9)
k=1
A B = C,
T 1 |
(3.4.6)
(3.4.7)
j=1
The matrix multiplication is the combination or the composition of two linear mappings
y=Ax
y = A (B z) = (A B) z = C z.
(3.4.8)
x=Bz
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
det T 6= 0.
det T 6= 0.
(3.4.10)
(3.4.11)
m
X
, and A Rnn .
(3.4.4)
!
m
!
m
m
n
n
n
X
X
X
X
X
X
yi =
(Aij xj ) =
Bjk zk =
Aij Bjk zk =
Cik zk .
Aij
j=1
x, y Rn
The two vectors x and y are described by the same linear mapping and the same nonsingular
and y. The vectors are called to be similar, because they are
square matrix T and the vectors x
transformed in the same way
k=1
j=1
, with
T y = A T x
,
(3.4.12)
,
y = T 1 A T x
(3.4.13)
A = T 1 A T .
(3.4.14)
and nally
The matrix A = T 1 A T is the result of the so called similarity transformation of the matrix
A with the nonsingular transformation matrix T . The matrices A and A are said to be similar
matrices.
56
the determinants are equal
57
det A = det T
= det T
AT ,
det A det T
det T
1
=
,
det T
det A = det A.
condition:
(3.4.15)
(3.4.23)
assumption: A = A ,
tr A = tr T 1 A T ,
= tr A T T 1 ,
(3.4.17)
(3.4.25)
assumption: P = xT y = P = x
T y,
(
x
= T 1 x
proof:
x=T x
y = T T y
T
x
T y = T 1 x T T y
T
= xT T 1 T T y
1 T
= xT T T
T y
(3.4.26)
, with
det T 6= 0.
(3.4.27)
= xT y
, with
x, y Rn
, and A Rnn .
(3.4.18)
with a square matrix A. The vectors x and y are computed in an opposite way (kontragredient)
with the nonsingular square matrix T and the vectors x
and y
, with
x=T x
(3.4.24)
A = (T A T ) = T A T = T A T = A.
proof:
(3.4.16)
tr A = tr A.
y=Ax
x, x
Rn
T Rnn
, and
det T 6= 0.
(3.4.19)
The y is the result of the mutliplication of the transpose of the matrix T and the vector y
y = T T y
, with
y, y Rn
T Rnn
, and
det T 6= 0.
(3.4.20)
T |
y=AT x
,
y=Ax
x, y Rn
, and A Rnn .
(3.4.28)
The vectors x and y will be transformed in the similar way and in the congruent way with the so
called orthogonal matrix T = Q, det Q 6= 0.
x=Qx
,
and nally
,
y = A x
(3.4.21)
A = TTA T.
(3.4.22)
The matrix product A = T T A T is called the congruence transformation of the matrix A. The
matrices A and A are called to be congruent matrices.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
, with
y = Q y
y = Q1 y
T
y = Q y
similar transformation,
(3.4.29)
congruent transformation.
(3.4.30)
For the orthogonal transformation the transformations matrices are called to be orthogonal, if
they fulll the relations
(3.4.31)
Q1 = QT or Q QT = 1.
For the orthogonal matrices the following identities hold.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
58
59
and
y = QT y,
cos sin y1
y1
=
y2
sin cos y2
(3.4.32)
(3.4.35)
y = QT y.
(3.4.36)
y = Q1 y
Q1 := x
; Xik =
ki
Q
.
det Q
(3.4.37)
Solving this equations step by step, starting with computing the determinant,
y2 sin
y1 cos
:9
O9
y2 -axis y2 -axis 6
ki = (1)k+i det Q ,
Q
ki
y2
y2 cos
y1 sin
?
6
(3.4.38)
y2
y1 sin
y2 cos
y1 -axis
y1
O
W
y1
X11 =
(3.4.40)
X12
Q22
= cos ,
1
= (1)3 Q12 = (1)3 ( sin ) = + sin ,
(3.4.41)
(3.4.42)
(3.4.43)
and nally
y1 -axis
y2 sin
y1 cos
(3.4.39)
X = Q1 =
cos sin
.
sin cos
(3.4.44)
Q1 = QT .
The most important usage of this rotation matrices is the rotation transformation of coordinates.
For example the rotation transformation in R2 is given by
y = Q y,
cos sin y1
y1
=
sin cos
y2
y2
(3.4.33)
y = Q y,
(3.4.34)
(3.4.45)
, with
B Rnn
, AT Rnm
, and A Rmn .
(3.4.46)
60
The matrix B is symmetric, i.e.
61
For example
B = BT ,
(3.4.47)
B T = (AT A)T = AT A = B.
(3.4.48)
because
If the rows and columns A are nonsingular, then the matrix B is nonsingular, i.e. the determinant
is not equal to zero,
det B 6= 0.
(3.4.49)
This product was introduced by Gauss in order to compute the so called normal equation. The
matrix A is given by
a
a
a
A=
n
1 2
2 1
13 4
A = 0 3 ; AT A =
,
4 14
3 2
sX
p
N (A) = 22 + 12 + 32 + (3)2 + 22 = 3 3 =
A2ik ,
i,k
13 + 14 = 3 3 = tr AT A .
(3.4.50)
T
an
T
a1 a1 aT1 a2
T
a a aT a
= 2. 1 2. 2
..
..
aTn a1
(3.4.51)
an
,
aT1 an
..
.
.
..
...
.
T
an an nn
(3.4.52)
An element Bik of the product matrix is the scalar product of the i-th column vector with the k-th
column vector of A,
Bik = aTi ak .
(3.4.53)
The diagonal elements are called the quadratic value of the norm of the column vectors and this
value is always positive (ai 6= 0). The sum, i.e. the trace of the product AT A or the sum of all
A2ik , is the quadratic valued of a matrix norm, called the Euklidian matrix norm N (A),
N (A) =
sX
q
tr(AT A) =
A2ik .
(3.4.54)
i,k
62
3.5
Quadratic Forms
63
= y T T T A T y,
x Rn
y Rn
, and
Aik R,
(3.5.1)
det A 6= 0 , and A = A .
(3.5.2)
:= xT y = xT A x.
(3.5.3)
the product
is a real number, R, and is called the quadratic form of A. The following conditions hold
= T
=x Ax= =x A x
, because A = A ,
(3.5.11)
B = TTA T,
(3.5.12)
where T is a real nonsingular matrix. Than the matrices B and A are called to be congruent to
each other,
c
A B.
(3.5.13)
The congruence transformation bewares the symmetry of the matrix A, because the following
equation holds,
(3.5.14)
B = BT .
(3.5.4)
(3.5.5)
i.e. the matrix A must be symmetric. The scalar quantity and than the matrix A, too, are called
to be positive denite (or negative denite), if the following conditions hold,
= 0 , iff x = 0
It is necessary, that the determinant does not equal zero det A 6= 0, i.e. the matrix A must be
nonsingular. If there exists a vector x 6= 0, such that = 0, then the form = xT A x is called
semidenite. In this case if the matrix A is singular, i.e. det A = 0, then the homogenous system
of equations,
(3.5.7)
Ax=0
xT A x = 0 , iff x 6= 0 , and det A = 0 ,
or resp.
a1 x1 + a2 x2 + . . . + an xn = 0,
= y T B y.
(3.5.10)
(3.5.8)
, and AT = A.
(3.5.15)
should be partial derived w.r.t. the components of the vector x. The result forms the column
,
matrix
x
0
0
.
.
x
.
= = ei , i-th unit vector,
(3.5.16)
xi 1
.
..
0
and
xT
= 0 0 1 0 = eTi .
xi
With equations (3.5.16) and (3.5.17) the derivative of the quadratic form is given by,
has got only nontrivial solutions, because of the linear dependence of the columns of the matrix
A. The condition xT A x = 0 could only hold, iff the vector is nonequal to zero, x 6= 0, and the
determinant of the matrix A equals zero, det A = 0.
(3.5.17)
= eTi A x + xT A ei .
xi
(3.5.18)
A = AT ,
(3.5.19)
, and AT = A.
(3.5.9)
T
T
x A ei = eTi AT x = eTi A x,
(3.5.20)
64
3.6
and nally
= 2 eTi A x.
xi
(3.5.21)
The quantity eTi A x is the i-th component of the vector A x. Furthermore the n derivatives
are combined as a column matrix
1 0 0
x1
x2
0 1 0
= . = 2 .. .. . . .. A x
x ..
. .
. .
1
xn
xi
= 21 A x
= 2A x.
x
(3.5.22)
65
, with
x, y Rn
xi , yi , Aik F
, and
det A 6= 0,
(3.6.1)
, and F.
(3.6.2)
The directions associated to the eigenvector x0 is called a principal axis. The whole task is
described as the principal axes problem of the matrix A. The scalar quantity is called the
eigenvalue, because of this denition the whole problem is also called the eigenvalue problem.
The equation (3.6.2) could be rewritten like this
y 0 = 1 x0 ,
(3.6.3)
(3.6.4)
(3.6.5)
The so called special eigenvalue problem is characterized by the eigenvalues distributed only
on the main diagonal. For the homogeneous linear equation system of the x0 exists a trivial
solution x0 = 0. A nontrivial solution exists only if this condition is fullled,
det (A 1) = 0.
(3.6.6)
This equation is called the characteristic equation, and the left-hand side det (A 1) is called
the characteristic polynomial. The components of the vector x0 are yet unknown. The vector
x0 could be computed by determing the norm, because the principal axes are searched. Solving
the determinant implies for a matrix with n rows a polynomial of n-th degree. The roots or
sometimes also called the zeros of this equation or polynomial are the eigenvalues.
p () = det (1 A) = n + an1 n1 + . . . + a1 + a0 .
(3.6.7)
The rst and the last coefcient of the polynomial are given by
an1 = tr A, and
a0 (1)n = det A.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.6.8)
(3.6.9)
66
With the polynomial factorization the equation p () or the polynomial (3.6.7) could be described
by
(3.6.10)
p () = ( 1 ) ( 2 ) . . . ( n ) .
Comparing this with Newtons relation for a symmetric polynomial the equations (3.6.8) and
(3.6.9) could be rewritten like this,
tr A = 1 + 2 + . . . + n , and
det A = 1 2 . . . n .
(3.6.11)
(3.6.12)
Because the associated eigenvectors to each eigenvalue could not be computed explict, the whole
equation is normed
y 0i = A xi0 = i x0i
, with i = 1, 2, 3, . . . , n,
(3.6.13)
and the eigenvectors x0i . If for example the matrix (A 1) has the reduction of rank d = 1 and
(1)
the vector x0i is an arbitrary solution of the eigenvalue problem, then the equation,
(1)
x0i = cx0i ,
results
(1 2 ) xT1 x2 = 0,
(3.6.20)
, and
x1 = b + ic , and
3.6.1.0.15 1st Rule. The eigenvectors x0i of a nonsingular and symmetric matrix A are orthogonal to each other.
Proof. Let the vectors x0i = xi , x1 and x2 be eigenvectors
2 = i,
(3.6.22)
x2 = b ic.
(3.6.23)
implies
(3.6.15)
and nally
(3.6.16)
xT1 (A 2 1) x2 = 0,
(3.6.17)
and also
(1 2 ) xT1 x2 = 0,
(3.6.24)
(1 2 ) bT + icT bT icT = 0,
(3.6.25)
2i bT b + cT c = 0.
(3.6.26)
T
T
This equation implies = 0, because the term b b + c c 6= 0 is nonzero, i.e. the eigenvalues
are real numbers.
(3.6.19)
and because (1 2 ) 6= 0, and the scalar product xT1 x2 = 0 must equal zero, this means that
the vectors are orthogonal
x1 x2 , iff 1 6= 2 .
(3.6.21)
A = AT ,
(A 1 1) x1 = 0,
, and
(3.6.14)
with the parameter c represents the general solution of the eigenvalue problem. If the reduction
of rank of the matrix is larger than 1, then there exist d > 1 linearly independent eigenvectors
x1 . As a rule of thumb,
3.6.1.0.16
67
(3.6.18)
(A 1) x = 0
, with
A = AT
det A 6= 0
, Aij R,
(3.6.27)
68
69
, with
R.
(3.6.28)
For large matrices A the setting-up and the solution of the characteristic equation is very complicated. Furthermore for some problems it is sufcient to know only the largest and/or the smallest
eigenvalue, e.g. for a stability problem only the smallest eigenvalue is of interest, because this is
the critical load. Therefore the so called direct method to compute the approximated eignevalues
1 by von Mises is interesting. For using this method it is necessary to compute the inverse before starting with the actual method to determine the smallest, critical load case. This so called
von Mises iteration is given by
z = A z 1 = A z 0 .
(3.6.29)
In this iterative process the vector z converges to x1 , i.e. the vector converges to the eigenvalue 1 with the largest absolute value. The starting vector z 0 is represented by the linearly
independent eigenvectors xi ,
z 0 = C1 x1 + C2 x2 + . . . + Cn xn 6= 0,
(3.6.30)
(3.6.36)
(3.6.37)
with the matrices A and B being nonsingular. The eigenvalues are multiplied with an arbitrary
matrix B and not with the identity matrix 1. This problem is reduced to the special eigenvalue
problem by multiplication with the inverse of matrix B form the left-hand side,
1
B A 1 x = 0,
(3.6.38)
1
(3.6.39)
B A 1 x = 0.
Even if the matrices A and B are symmetric, the matrix C = B 1 A is in general a nonsymmetric
matrix, because the matrix multiplication is noncommutative.
(3.6.31)
(3.6.32)
(3.6.33)
()
qi
zi
(1)
zi
1 .
(3.6.34)
The convergence will be better, if the ratio |1 | / |2 | increases. A very good approximated value
1 for the dominant (largest) eigenvalue 1 is established with the so called Rayleigh quotient,
zT z
zT A z
1 = R [z ] = T +1 = T
z z
z z
, with
1 1 .
(3.6.35)
The numerator and the denominator of the Rayleigh quotient include scalar products of the ap()
proximated vectors. For this reason the information of all components q i are used in this
approximation.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.6.40)
If the condition |1 | |2 | |3 | . . . n holds, then with the raising value of the vector
z u converges to the eigenvector x1 multiplied with a constant c1 ,
z 1 c1 x1 ,
z +1 1 z .
A x = y = x = 1 x,
, and y = T y
x=T x
, with
y = 1 x
.
(3.6.41)
1
= 0, and
T A T 1 x
1 x
= 0.
A
(3.6.42)
(3.6.43)
det T 1 =
1
,
det T
(3.6.44)
(3.6.45)
70
71
3.6.4.0.19 Rule. The eigenvalues of the matrix A do not change if the matrix is transformed
1 = det (A 1) = 0.
det T 1 A T 1 = det A
(3.6.46)
The nonsingular symmetric matrix A with n rows contains n linearly independent eigenvectors xi , if and only if for any multiple eigenvalue (, i.e. multiple roots of the characteristic
polynomial) of multiplicity p the reduction of rank d = p (, with = 1, 2, . . . , s) for the
characteristic matrix (A 1) equals the multiplicity of the multiple eigenvalue. The quantity s
describes the number of different eigenvalues. The n linearly independent normed eigenvectors
xi of the matrix A are combined as column vectors to form the nonsingular eigenvector matrix,
The Cayley-Hamilton Theorem says, that an arbitrary square matrix A satises its own characteristic equation. If the characteristic polynomial for the matrix A is
p () = det (1 A) ,
= n + an1 n1 + . . . + a1 + a0 ,
X = [x1 , x2 , . . . , xn ]
, with
det X 6= 0.
(3.6.47)
1 0 0
..
.
0 2
[1 x1 , . . . , n ] = [x1 , x2 , . . . , xn ] .
,
. . . ..
..
.
0 n
[1 x1 , . . . , n ] = X .
(3.6.48)
(3.6.49)
(3.6.50)
(3.6.51)
(3.6.52)
and nally
X 1 A X = X , with
det X 6= 0.
(3.6.54)
or
AT =T D
, with
T = [t1 , . . . , tn ] ,
(3.6.55)
and nally
A ti = Dii ti .
(3.6.56)
The column vectors ti of the transformation matrix T are the n linearly independent eigenvectors
of the matrix A with the associated eigenvalues i = Dii .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.6.59)
(3.6.60)
The matrix with the exponent n, written like A , could be described by a linear combination
of the matrices with the exponents n 1 up to 0, resp. An1 till A0 = 1. If the matrix A is
nonsingular, then also negative quantities as exponents are allowed, e.g.
A1 = p (A) = 0,
a0
n1
+ an1 A
n2
(3.6.61)
+ . . . + a1 = 0
, a0 6= 0.
(3.6.62)
Furthermore the power series P (A) of a matrix A, with the eigenvalues appearing -times
in the minimal polynomial, converges, if and only if the usual power series converges for all
eigenvalues of the matrix A. For example
A
1
1
e = 1 + A + A 2 + A3 + . . . ,
(3.6.63)
2!
3!
1 2 1 4
(3.6.64)
[cos (A)] = 1 A + A + . . . ,
2!
4!
1 3 1 5
[sin (A)] = A A + A + . . . .
(3.6.65)
3!
5!
(3.6.53)
Therefore the diagonal matrix of eigenvalues could be computed by the similarity transformation of the matrix A with the eigenvector matrix X. In the opposite direction a transformation
matrix T must fulll some conditions, in order to transform a matrix A by a similarity transformation into a diagonal matrix, i.e.
T 1 A T = D = dDi ic,
(3.6.57)
(3.6.58)
(3.6.66)
with the ci called the evaluation coefcients. Introducing some basic vectors and matrices, in
order to establish the evaluation theorem,
X = [x1 , x2 , . . . , xn ] ,
(3.6.67)
c = [c1 , c2 , . . . , cn ]T ,
X c = z, and
(3.6.68)
(3.6.69)
c = X 1 z.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.6.70)
72
Let z 0 be an arbitrary real vector to start with, and establish some iterated vectors
(3.6.71)
z 2 = A z 1 = A2 z 0 ,
...
z n = A z n1 = An z 0 .
(3.6.72)
(3.6.73)
The n + 1 vectors z 0 till z n are linearly dependent, because every n + 1 vectors in Rn must be
linearly dependent. The characteristic polynomial of the matrix A is given by
(3.6.74)
(3.6.75)
The relation between the starting vector z 0 and the rst n iterated vectors z i is given by the
following equations, the evaluation theorem
z 0 = c 1 x 1 + c 2 x2 + . . . + c n xn ,
(3.6.76)
, and
p () = det (1 A) = 0.
73
z1 = A z0,
p () = det (1 A) ,
= a0 + a1 + . . . + an1 n1 + n .
(3.6.77)
a0 z 0 + a1 z 1 + . . . + z n = a0 + a1 1 + . . . + an1 1n1 + n1 c1 x1
+ a0 + a1 2 + . . . + an1 2n1 + n2 c2 x2
..
.
+ a0 + a1 n + . . . + an1 nn1 + nn cn xn .
(3.6.82)
(a0 1 + a1 A + . . . + An ) z 0 = p (1 ) c1 x1 + p (2 ) c2 x2 + . . . + p (n ) cn xn ,
p (A) z 0 = 0 c1 x1 + 0 c2 x2 + . . . + 0 cn xn ,
and nally
p (A) z 0 = a0 z 0 + a1 z 1 + . . . + z n = 0.
(3.6.83)
(3.6.84)
(3.6.85)
(3.6.78)
and with an arbitrary vector z 0 the term in brackets must equal the zero matrix,
a0 1 + a1 A + . . . + An = 0.
and
z 1 = A; z 0 ,
z 1 = c1 A; x1 + c2 A; x2 + . . . + cn A; xn ,
z 1 = 1 c 1 x1 + 2 c 2 x2 + . . . + n c n xn ,
..
.
z n = n1 c1 x1 + n2 c2 x2 + . . . + nn cn xn ,
In other words, an arbitrary square matrix A solves its own characteristic equation. If the characteristic polynomial of the matrix A is given by equation (3.6.74), then the matrix A solves the
so called Cayley-Hamilton equation,
(3.6.79)
(3.6.80)
(3.6.86)
p (A) = a0 1 + a1 A + . . . + An = 0.
(3.6.87)
The polynomial p (A) of the matrix A equals the zero matrix, and the a i are the coefcients of
the characteristic polynomial of matrix A,
p () = det (1 A) = n + an1 n1 + . . . + a1 + a0 = 0.
(3.6.81)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(3.6.88)
74
Chapter 4
Vector and Tensor Algebra
For example S IMMONDS [12], H ALMOS [6], M ATTHEWS [11], and A BRAHAM, M ARSDEN, and
R ATIU [1].
And in german DE B OER [3], S TEIN ET AL . [13], and I BEN [7].
75
76
4.2
4.1.4
4.1.5
4.1.6
4.1.7
4.2.3
4.2.5
4.2.6
4.2.7
4.2.8
4.2.9
Tensors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
4.3.1
4.3.2
4.3.3
4.5.7
4.5.8
4.5.9
4.7
4.6.2
4.6.3
4.6.4
4.6.5
4.6.6
4.6.7
Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
4.6.8
4.7.2
4.7.3
4.7.4
4.7.5
4.5
4.6
4.2.4
Products of Vectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
4.2.2
4.4
4.1.3
4.2.1
4.3
4.5.5
4.5.6
4.4.2
4.4.3
4.4.4
4.5.2
4.5.3
4.5.4
77
78
4.1
79
If the indices of two terms are written in brackets it is forbidden to sum these terms
v(m) g(m) 6=
(4.1.6)
m=1
(4.1.1)
ij = ji = ij = ij =
v gj = v g1 + v g2 + v g3 = v gj ,
(4.1.2)
ij vj =
1
0
if
if
i=j
.
i 6= j
(4.1.7)
An index i, for example in a 3-dimensional space, is substituted with another index j by multiplication with the Kronecker delta,
j=1
3 X
3
X
v m gm .
i=1
3
X
3
X
3
X
ij vj = i1 v1 + i2 v2 + i3 v3 = vi ,
(4.1.8)
j=1
i=1 k=1
(4.1.3)
The repeated index of summation is also called the dummy index. This means that changing the
index i to j or k or any other symbol does not infect the value of the sum. But is important to
notice, that it is not allowed to repeat an index more than twice! Another important thing to note
about index notation is the use of the free indices. The free indices in every term and on both
sides of an equation must match. For that reason the addition of two vectors could be written in
different ways, where a, b and c are vectors in the vector space V with the dimension n, and the
ai , bi and ci are their components,
a1 + b 1 = c 1
a 2 + b 2 = c 2
, a, b, c V.
(4.1.4)
a + b = c a i + bi = c i
...
a + b = c
n
For the special case of Cartesian coordinates there holds another important convention. In this
case it is allowed to sum repeated subscript or superscript indices, in general for a Cartesian
coordinate system the subscript index is preferred,
3
X
xi ei = x i e i .
i=1
(4.1.5)
ij v i uj =
3 X
3
X
ij v i uj
i=1 j=1
= 11 v 1 u1 + 12 v 1 u2 + 13 v 1 u3
+ 21 v 2 u1 + 22 v 2 u2 + 23 v 2 u3
+ 31 v 3 u1 + 32 v 3 u2 + 33 v 3 u3
= 1 v 1 u1 + 0 v 1 u2 + 0 v 1 u3
+ 0 v 2 u1 + 1 v 2 u2 + 0 v 2 u3
+ 0 v 3 u1 + 0 v 3 u2 + 1 v 3 u3
ij v i uj = v 1 u1 + v 2 u2 + v 3 u3 = v i ui
v u,
(4.1.9)
3
X
jj = 11 + 22 + 33 = 3,
(4.1.10)
j=1
3
X
ij jk = i1 1k + i2 1k + i3 3k = ik .
j=1
(4.1.11)
80
For the special case of Cartesian coordinates the Kronecker delta is identied by the unit matrix
or identity matrix,
1 0 0
(4.1.12)
[ij ] = 0 1 0 .
0 0 1
, with
v, gi V , and
i = 1, 2, 3.
(4.1.13)
The vectors gi are choosen as linear independent, i.e. they are a basis. If the index i is an
subscript index, the denitions
gi
81
in the same space like the covariant base vectors. This contravariant base vectors are dened by
(
1 i=k
k
k
gi g = i =
,
(4.1.21)
0 i 6= k
and with the covariant coordinates vi the vector v is given by
v = vi g i
, with
v, gi V , and i = 1, . . . , n.
(4.1.14)
g2 O
and
vi
, contravariant coordinates,
g2
(4.1.15)
of v with respect to the gi , hold. The v 1 g1 , v 2 g2 , v 3 g3 are called the components of v. The
Scalar product of the base vectors gi and gk is dened by
gi gk = gik
= gk gi = gki
gik = gki
(4.1.16)
(4.1.17)
(4.1.18)
g1
g1
(4.1.22)
The metric coefcients are symmetric, because of the commutativity of the scalar product g i
gk = gk gi . The determinant of the matrix of the covariant metric coefcients g ik ,
g = det [gik ]
(4.1.19)
is nonzero, if and only if the gi form a basis. For the Cartesian basis the metric coefcients
vanish except the ones for i = k and the coefcient matrix becomes the identity matrix or the
Kronecker delta
(
1 i=k
ei ek = ik =
.
(4.1.20)
0 i 6= k
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
g1 g 2
g2 g 1
g1 g 1
g2 g 2
= 0 g1 g2 ,
= 0 g2 g1 ,
= 1,
= 1.
(4.1.23)
(4.1.24)
(4.1.25)
(4.1.26)
= gk gi = g ki
g =g
ki
(4.1.27)
(4.1.28)
(4.1.29)
82
83
Lemma 4.1. The covariant base vectors transform with the contravariant metric coefcients
into the contravariant base vectors.
e2 = e 2 6
gk = g ki gi
The same argumentation for the covariant metric coefcients starts with
e1 = e 1
gk = Akm gm ,
gk gi = Akm im ,
gki = Aki ,
(4.1.36)
(4.1.37)
(4.1.38)
gk = gki gi .
(4.1.39)
As a rule of thumb:
Lemma 4.2. The contravariant base vectors transform with the covariant metric coefcients
into the covariant base vectors.
e3 = e 3
gk = gki gi
denes the
g ik = g ki
For the special case of Cartesian coordinates and an orthonormal basis e i co- and contravariant
base vectors are equal. For that reason it is not necessary to differentiate between indices as
subscript or superscript indices. From now on Cartesian base vectors and Cartesian coordinates
get only indicies as subscript indices,
u = u i ei
, or u = uj ej .
(4.1.30)
(4.1.40)
gk gi = g km gm gi .
(4.1.41)
Comparing this with the denitions of the Kronecker delta (4.1.7) and of the metric coefcients
(4.1.16) and (4.1.27) leads to
ik = g km gmi .
If the vectors gi , gm and gk are in the same space V, it must be possible to describe g k by a
product of gm and some coefcient like A km ,
gk = Akm gm
gk = g km gm ,
(4.1.31)
(4.1.32)
(4.1.33)
g ki = Aki .
(4.1.34)
The result is the following relation between co- and contravariant base vectors
gk = g ki gi .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.1.35)
(4.1.42)
Like in the expression A A = 1 co- und contravariant metric coefcients are inverse to each
other. In matrix notation equation (4.1.42) denotes
1 = g km [gmi ] ,
(4.1.43)
km
1
= [gmi ] ,
(4.1.44)
g
and for the determinants
1
.
(4.1.45)
det [gik ]
With the denition of the determinant, equation (4.1.19), the determinant of the contravariant
metric coefcients gives
(4.1.46)
det [gik ] = g,
det g ik =
(4.1.47)
84
4.2
v i gik gk = vk gk
vk = gik v i .
vk = gki v i .
(4.1.48)
(4.1.49)
In the same way comparing the contravariant vector g k = g ik gi with the equations (4.1.35) and
(4.1.18) gives
(4.1.50)
v i gi = vk g ki gi v i = g ki vk ,
and after changing the indices
ik
v = g vk .
(4.1.51)
Lemma 4.3. The covariant coordinates transform like the covariant base vectors with the contravariant metric coefcients and vice versa. In index notation the transformation for the covariant coordinates and the covariant base vectors looks like this
v i = g ik vk
gi = g ik gk ,
(4.1.52)
Products of Vectors
gk = gki gi .
u = u i gi
and
(4.2.1)
v = v i gi ,
(4.2.2)
and v = vi gi .
(4.2.3)
w.r.t. the contravariant base vectors g j and j = 1, . . . , n. By combining these representations the
scalar product of two vectors could be written in four variations
= u v = ui v j gi gj = ui v j gij = ui vi ,
= ui vj gi gj = ui vj g ij = ui v i ,
= ui vj gi gj = ui vj i.j = ui vi ,
= ui v g g j =
(4.1.53)
, R,
and also called the inner product or dot product of vectors. The vectors u and v are represented
by
j i
and for the contravariant coordinates and the contravariant base vectors
vk = gki v i
85
After changing the indices of the symmetric covariant metric coefcient, like in equation (4.1.29),
the transformation from contravariant coordinates to covariant coordinates denotes like this
ui v j .ji
= ui v .
(4.2.4)
(4.2.5)
(4.2.6)
(4.2.7)
The Euclidean norm is the connection between elements of the same dimension in a vector space.
The absolute values of the vectors u and v are represented by
(4.2.8)
|u| = kuk2 = u u,
|v| = kvk2 = v v.
(4.2.9)
The scalar product or inner product of two vectors in V is a bilinear mapping from two vectors
to R.
Theorem 4.1. In the 3-dimensional Euclidean vector space E 3 one important application of the
scalar product is the denition of the work as the force times the distance moved in the direction
opposite to the force,
Work = Force in direction of the distance Distance
or = f d .
(4.2.10)
Theorem 4.2. The scalar product in 3-dimensional Euclidean vector space E 3 is written u v
and is dened as the product of the absolute values of the two vectors and the cosine of the angle
between them,
= u v := |u| |v| cos .
(4.2.11)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
86
87
vf
:
eu
ud
|v| cos
The cross product, also called the vector product, or the outer product, is only dened in the
3-dimensional Euclidean vector space E3 . The cross product of two arbitrary, linear independent
covariant base vectors gi , gj E3 implies another vector gk E3 and is introduced by
The quantity |v| cos represents in the 3-dimensional Euclidean vector space E the projection
of the vector v in the direction of vector u. The unit vector in direction of vector u is given by
gi gj = gk ,
eu =
u
.
|u|
(4.2.12)
uv
.
|u| |v|
(4.2.13)
(4.2.18)
i, j, k = 1, 2, 3
i 6= j 6= k,
, or another even permutation of i, j, k.
The cross products of the Cartesian base vectors ei in the 3-dimensional Euclidean vector space
E3 are given by
u u , and
|v| =
v v.
This formula rewritten with the base vectors gi and gi simplies in index notation to
q
p
p
|u| = ui gi uk gk = ui uk ik = ui ui ,
p
|v| = v i vi .
(4.2.14)
(4.2.15)
(4.2.16)
The cosine between two vectors in the 3-dimensional Euclidean vector space E 3 is dened by
i
cos = p
u vi
ui v
=p
.
uj uj v k v k
uj uj v k v k
(4.2.17)
For example the scalar product of two vectors w.r.t. the Cartesian basis g i = ei = ei in the
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
e1 e 2
e2 e 3
e3 e 1
e2 e 1
e3 e 2
e1 e 3
= e3 = e3 ,
= e1 = e1 ,
= e2 = e2 ,
= e3 = e3 ,
= e1 = e1 ,
= e2 = e2 .
(4.2.19)
(4.2.20)
(4.2.21)
(4.2.22)
(4.2.23)
(4.2.24)
The Cartesian components of a permutation tensor , or just the permutation symbols, are dened
by
88
Thus, returning to equations (4.2.19)-(4.2.24), the cross products of the Cartesian base vectors
could be described by the permutation symbols like this,
k
ei ej = eijk e .
(4.2.26)
For example
i, j, k = 1, 2, 3
(4.2.27)
(gi gj ) gk = g gk ,
[g1 , g2 , g3 ] = kk = 3.
(4.2.28)
(4.2.29)
This result is the so called scalar triple product of the base vectors
= [g1 , g2 , g3 ] .
(4.2.30)
gi g j
,
[g1 , g2 , g3 ]
(4.2.31)
gi gj
.
[g1 , g2 , g3 ]
(4.2.32)
a
a
a
a
a
a
= a11 22 23 a12 21 23 + a13 21 22
a32 a33
a31 a33
a31 a32
= a11 a22 a33 a11 a32 a23
a12 a21 a33 + a12 a31 a23
+ a13 a21 a32 a13 a31 a22 .
(4.2.36)
Furthermore the scalar product of two scalar triple products of base vectors is given by
1
[g1 , g2 , g3 ] g1 , g2 , g3 = = 1.
The scalar quantity in the section above could also be described by the square root of the
determinant of the covariant metric coefcients,
1
= (det gij ) 2 = g.
(4.2.34)
This scalar triple product of the base vectors gi for i = 1, 2, 3 represents the volume of the
parallelepiped formed by the three vectors gi for i = 1, 2, 3. Comparing equations (4.2.28) and
(4.2.29) implies for contravariant base vectors
gk =
and nally
The gk are the contravariant base vectors and the scalar quantity is computed by multiplication
of equation (4.2.18) with the covariant base vector gk ,
k
89
gi gj = gk ,
i 6= j 6= k,
, or another even permutation of i, j, k.
(4.2.37)
(4.2.38)
90
The matrix of the covariant metric coefcients is the inverse of the matrix of the contravariant
metric coefcients and vice versa,
det gij g
The product rule of determinants
jk
gij g jk = ik ,
= det [gij ] g jk = det ik = 1.
1
1=g ,
g
and nally
det [gij ] = g
1
det g ij = .
g
(4.2.40)
[d, e, f ] = (d e) f .
(4.2.51)
1 2 3 a1 a2 a3
= g , g , g b b b
c1 c2 c3
1 1 1
1 1 1
a b c
a b c
2 2 2
= [g1 , g2 , g3 ] a b c = a2 b2 c2 .
3
3
3
3
3
3
a b c
a b c
(4.2.41)
(4.2.42)
(4.2.43)
(4.2.44)
(4.2.45)
The vectors a up to f are written in the 3-dimensional Euclidean vector space E 3 with the base
vectors gi and gi ,
a = a i gi
d = d i gi ,
(4.2.46)
b = b i gi
e = e i gi ,
(4.2.47)
c = c i gi
f = f i gi .
(4.2.48)
The cross product (4.2.26) rewritten with the formulae for the scalar triple product (4.2.28) (4.2.30),
a b = ai gi bj gj = ai bj eijk [g1 , g2 , g3 ] gk ,
1 2 3
1 2 3
a a a
g g g
1 2 3
a b = [g1 , g2 , g3 ] b b b = [g1 , g2 , g3 ] a1 a2 a3 .
g1 g2 g3
b1 b2 b 3
(4.2.49)
[a, b, c] = (a b) c,
(4.2.50)
(4.2.52)
The same formula written with covariant components and contravariant base vectors,
The product
1
(a b) c = g1 , g2 , g3 ai bj ck eijk = ai bj ck eijk
1 2 3 a 1 a2 a 3
= g , g , g b1 b2 b3
c1 c2 c3
1 1 1
a b c
= [g1 , g2 , g3 ] a2 b2 c2 .
a3 b 3 c 3
P = [a, b, c] [d, e, f ]
1 2 3
1 2 3 a1 a2 a3 d1 e1 f1
1
P = [g1 , g2 , g3 ] g , g , g b b b d2 e2 f2 = |A| |B| = |A| |B| .
c1 c2 c3 d 3 e 3 f3
(4.2.53)
(4.2.54)
(4.2.55)
The element (1, 1) of the product matrix A B with respect to the product rule of determinants
det A det B = det (A B) is given by
a1 d 1 + a 2 d 2 + a 3 d 3 = a i g i d j g j
= ai dj ij
91
and
(4.2.39)
For this reason the determinants of the matrix of the metric coefcients are represented with the
permutation symbols, see equations (4.2.35) and (4.2.43), like this
ijk
= ai di
= a d.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.2.56)
92
Comparing this with the product P leads to
a d a e a f
P = [a, b, c] [d, e, f ] = b d b e b f ,
c d c e c f
and for the scalar triple product [a, b, c] to the power two,
a a a b a c
2
[a, b, c] = b a b b b c .
c a c b c c
(4.2.57)
Comparing equations (4.2.25) with (4.2.65) and (4.2.66) shows the relations,
ijk =
The square value of a scalar triple product of the covariant base vectors, like in equations (4.2.58),
g1 g 1 g 1 g 2 g1 g 3
g.
gi gj = ijk gk ,
(4.2.63)
g i g j = g 1 , g 2 , g 3 gk
1
= eijk gk ,
g
gi gj = ijk gk .
(4.2.64)
For example the general permutation symbol could be given by the covariant symbol,
(4.2.65)
ijk = g if (i, j, k) is an odd permutation of (1, 2, 3),
geijk
1
, and eijk = ijk ,
g
(4.2.67)
ijk
g .
(4.2.68)
and
1
ijk = eijk
g
, and eijk =
(4.2.60)
The same for relation for the scalar triple product of the contravariant base vectors leads to
1 2 3 2
1
g ,g ,g
(4.2.61)
= det g ij =
g
1 2 3
1
(4.2.62)
g ,g ,g = .
g
gi gj = [g1 , g2 , g3 ] gk
= g eijk gk ,
(4.2.66)
(4.2.58)
reduces to
93
(4.2.69)
(4.2.70)
(4.2.71)
(4.2.72)
lmn = g li g mj g nk ijk .
(4.2.73)
and
The covariant symbols are converted into the contravariant symbols with the contravariant
metric coefcients and vice versa. This transformation is the same as the one for tensors. The
conclusion is that the symbols are tensors! The relation between the e and symbols is written
as follows
g
(4.2.74)
eijk elmn = ijk lmn ,
g
eijk elmn = ijk lmn .
The relation between the permutation symbols and the Kronecker delta is given by
i i
l m ni
j
ijk
ijk
j
j
l m n = lmn = e elmn .
k k k
m
n
l
(4.2.75)
(4.2.76)
94
After expanding the determinant and setting k = n,
ijk
lmk =
j
li m
i j
m
l ,
(4.2.77)
and if i = l and j = m
95
The proof that the dyadic product is a tensor starts with the assumption, see equations (T4) and
(T5),
(a b) (u + v) = (a b) u + (a b) v.
(4.2.86)
(4.2.78)
T (u + v) = T (u) + T (v) .
ijk
ijk
ijk = e eijk =
2kk
= 6.
(4.2.79)
T (u + v) = (a b) (u + v) = a [b (u + v)]
= a [b u + b v]
= [a (b u)] + [a (b v)]
The dyadic product of two vectors a and b V denes a so called simple second order tensor
with rank = 1 in the tensor space V V over the vector space V by
T=ab
T V V .
, and
(4.2.80)
This tensor describes a linear mapping of the vector v V with the scalar product by
T v = (a b) v = a (b v) .
The vectors a and b are represented by the base vectors gi (covariant) and gj (contravariant)
(4.2.81)
The dyadic product a b could be represented by a matrix, for example with a, b R and
T R 3 R3 ,
a1
T
b1 b2 b3 13
T = a b = a2
(4.2.82)
a3 31
a 1 b 1 a1 b2 a1 b 3
= a2 b 1 a2 b2 a2 b 3 .
(4.2.83)
a3 b1 a3 b2 a3 b3 33
The rank of this mapping is rank = 1, i.e. det T
(33)
a = a i gi
a 1 b 1 a1 b 2
a 2 b 1 a2 b 2
a 3 b 1 a3 b 2
T = a b = ai bj gi gj = T ij gi gj ,
(4.2.88)
(4.2.89)
and by
gi g j
det T ij = 0 ,
(22)
r T ij = 1
, and
rank = 1.
(4.2.90)
3
P
a1 i=1 bi vi
a1 b 3
v1
3
P
a2 b 3 v 2 =
a
b
v
i i ,
2
i=1
a3 b 3
v3
3
P
a3
bi v i
(4.2.84)
a bT v = T v = a bT v .
w = T v = T ij gi gj v k gk
= T ij v k gi gj gk
= T ij v k gi jk ,
i=1
or
, and gi , gj V.
b = bj g j
= T (u) + T (v) .
(4.2.87)
w = T ij v j gi = wi gi .
(4.2.85)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.2.91)
96
4.3
4.3. Tensors
97
Tensors
x3 6
dF
(2)
n(1)
(2)
= t dA(2)
I
n(2)
6 (n)
(P )
dA(2)
dA(n)
x1
dA(3)
x2
n(3)
Remarks
The scalar product F n = Fn ; |n| = 1 projects F cos on the direction of n, and the
result is a scalar quantity.
The cross product rF = MA establishs a vector of momentum at a point A in the normal
direction of the plane rA , F and perpendicular to F, too.
The dyadic product a b = T assigns a second order tensor T to a pair of vectors a and
b.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
The spaces R3 and E3 are homeomorphic, i.e. for all vectors x R3 and v E3 the same
rules and axioms hold. For this reason it is sufcient to have a look at the vector space R 3 .
Also the spaces Rn , En and V are homeomorphic, but with n 6= 3 the usual cross product
will not hold. For this reason the following denitions are made for the general vector
space V, but most of the examples are given in the 3-dimensional Euclidean vector space
E3 . In this space the cross product holds, and this space is the visual space.
(4.3.1)
The denitions and rules for linear spaces in chapter (2.4), i.e. the axioms of vector space (S1)
up to (S8) are rewritten for tensors T V V.
98
u V
v V.
3 . Axiom of Second Order Tensors. The unit tensor 1 sends any vector into itself,
u V, 1 V V.
5 . Axiom of Second Order Tensors. If the vector u is multiplied by a scalar, then the linear
mapping is denoted by
T (u) = Tu ; u V, R .
(T5)
4.3.2.0.22 Linearity for the Tensors.
6 . Axiom of Second Order Tensors. The multiplication with the sum of tensors of the same
space is distributive,
; u V, T1 , T2 V V.
(T6)
7 . Axiom of Second Order Tensors. The multiplication of a tensor by a scalar is linear, like
in equation (T5) the multiplication of a vector by a scalar,
(T) u = T (u)
u V, R .
n
X
Ti =
i=1
i
and
det T 6= 0
n
X
i=1
ai b i ,
T = a bi = ai bi ,
(4.3.2)
(4.3.3)
If the vectors ai and bi are represented with the base vectors gi and gi V like this,
ai = aij gj
bi = bil gl ,
(4.3.4)
(4.3.5)
(4.3.6)
The dyadic product of the base vectors includes one co- and one contravariant base vector. The
mixed components T jl of the tensor in mixed formulation are written with one co- and one contravariant index, too,
det T jl 6= 0.
(4.3.7)
If the contravariant base vector is transformed with the metric coefcient ,
gl = g lk gk ,
T = T jl gj g lk gk ,
(T8)
(T9)
T=
(4.3.8)
but like in matrix calculus not commutative, i.e. T1 T2 6= T2 T1 . The "product"T1 T2 of the
tensors is also called a "composition"of the linear mappings T 1 , T2 .
9 . Axiom of Second Order Tensors. The inverse of a tensor T1 is dened by
v = T u u = T1 v,
The simple second order tensor T V V is dened as a linear combination of n dyads and its
rank is n,
(T7)
(T3)
4 . Axiom of Second Order Tensors. The multiplication by a tensor is distributive with respect
to vector addition,
T (u + v) = Tu + Tv ; u, v V.
(T4)
(T1 + T2 ) u = T1 u + T2 u
99
(T1)
This mapping is the same like the mapping of a vector with a quadratic matrix in a space with
Cartesian coordinates.
2 . Axiom of Second Order Tensors. The action of the zero tensor 0 on any vector u maps the
vector on the zero vector,
0 u = 0u = 0 ; u, 0 V.
(T2)
1 u = 1u = u ;
4.3. Tensors
(4.3.9)
T = T jl g lk gj gk ,
(4.3.10)
(4.3.11)
(4.3.12)
100
4.4
implies
T = T jl gjk gk gl ,
T = Tkl g g
(4.3.14)
(4.3.15)
v = v i gi = v i g i = v i g i = v i g i .
(4.3.16)
ij
= T v gi gjk = T vj gi ,
w = w i gi
wi = T ij vj .
(4.3.17)
w = T v = T i j g i gj v k g k
(4.3.18)
w=
and
Ti vk gi jk =
wi gi , and
(4.3.13)
101
T i vj g ,
w i = T i vj
The relation between the two covariant base vectors gi and gi could be written with a second
order tensor like
g i = A gi .
(4.4.2)
If this linear mapping exists, then the coefcients of the transformation tensor A are given by
gi = 1gi = gk gk gi
(4.4.3)
k
= g gi gk = Aki gk ,
gi = Aki gk
(4.3.19)
(4.4.1)
, and Aki = gk gi .
A = gk gi gk gi = Aki gk gi .
Insert equation (4.4.5) in (4.4.2), in order to get the transformation (4.4.4) again,
i
gm = Aki gk gi gm = Aki m
gk = Akm gk .
(4.4.4)
(4.4.5)
(4.4.6)
(4.4.7)
This existence results out of its linear independence. The "retransformation"tensor A is again
dened by the multiplication with the unit tensor 1
gi = 1gi = gk gk gi
(4.4.8)
k
k
= g gi g k = A i g k ,
k
gi = A i g k
, and A i = gk gi .
(4.4.9)
k
A = g k gi g k g i = A i g k g i .
(4.4.10)
102
Inserting equation (4.4.10) in (4.4.7) implies again the transformation relation (4.4.9),
k
k i
k
gm = A i g k g i g m = A i m
gk = A m gk .
(4.4.11)
The tensor A is the inverse to A and vice versa. This is a result of equation (4.4.7) ,
1
A |
gi = A g i ,
gi = g i .
(4.4.12)
(4.4.13)
A=A
A A = 1,
(4.4.14)
(4.4.15)
In index notation with equations (4.4.4) and (4.4.9) the relation between the "normal"and the
"overlined"coefcients of the transformation tensor is given by
m
gi = Aki gk = Aki A k gm
ij
ij
m j
Aki A k m
,
k j
A iA k .
| gj ,
(4.4.16)
(4.4.17)
The transformation of contravariant basis works in the same way. If in equations (4.4.3) or
(4.4.8) the metric tensor of covariant coefcients is used in stead of the identity tensor, then
another representation of the transformation tensor is described by
gm = 1gm = gik gi gk gi
(4.4.18)
= gik Akm gi ,
gm = Aim gi
, and
(4.4.19)
If the transformed covariant base vectors g m should be represented by the contravariant base
vectors gi , then the complete tensor of transformation is given by
A = (gi gk ) gi gk = Aik gi gk .
(4.4.20)
The inverse transformation tensor T is given by an equation developed in the same way like the
equations (4.4.16) and (4.4.8). This inverse tensor is denoted and dened by
A = A1 = (gi gk ) gi gk = Aik gi gk .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.4.21)
103
gi = 1gi = gk gk gi = gk gi gk = Aki gk .
(4.4.22)
With this relationship the transformed (overlined) covariant base vectors are represented by the
covariant base vectors
k
g i = g g i gk = A i gk ,
(4.4.24)
and the transformed (overlined) covariant base vectors are represented by the contravariant base
vectors,
gi = Agi
, and
k
gi = (gk gi ) g = Aki g .
(4.4.25)
(4.4.26)
The relation between the contravariant base vectors in both systems of coordinates is dened by
gi = 1gi = gk gk gi = gk gi gk = Bki gk .
(4.4.27)
With this relationship the transformed (overlined) contravariant base vectors are represented by
the contravariant base vectors
gi = Bgi , and B = (gk gm ) gk gm = Bkm gk gm ,
gi = gk gi gk = Bki gk ,
(4.4.29)
gi = Bgi , and B = gk gm gk gm = B km gk gm ,
gi = gk gi gk = B ki gk .
(4.4.31)
(4.4.28)
and the transformed (overlined) contravariant base vectors are represented by the covariant base
vectors,
(4.4.30)
The inverse relation gi gi , and gk gk representing the "retransformations"from the transformed (overlined) to the "normal"system of coordinates are given by the following equations.
The inverse transformation between the covariant base vectors of both systems of coordinates is
denoted and dened by
k
gi = 1gi = gk gk gi = gk gi gk = Ai gk .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.4.32)
104
With this relationship the covariant base vectors are represented by the transformed (overlined)
covariant base vectors,
k
(4.4.33)
gi = Agi , and A = gk gm gk gm = Am gk gm ,
k
k
gi = g g i g k = A i g k ,
(4.4.34)
and the covariant base vectors are represented by the transformed (overlined) contravariant base
vectors,
gi = Ag
A = (gm gk ) g g = Amk g g ,
, and
gi = (gk gi ) gk = Aki gk .
(4.4.35)
(4.4.36)
The inverse relation between the contravariant base vectors in both systems of coordinates is
dened by
i
gi = 1gi = gk gk gi = gk gi gk = B k gk .
(4.4.37)
i
gi = gk gi gk = B k gk ,
gm ,
(4.4.38)
There exist the following relations between the transformation tensors A and A,
AA = 1 , or
k
Ami A m = ik
AA = 1
, i.e.
A i Akm = ik
(4.4.42)
BB = 1
, i.e.
B m Bkm = ki .
v = T mk uk gm = v m gm .
; Bmk = A m ,
w = Sv
Bmk =
, with
w V , and S V V ,
(4.4.50)
and the linear mapping (4.4.48) the associative law for the linear mappings holds
w = Sv = S (T u) = (ST) u.
(4.4.51)
The second linear mapping w = Sv in the mixed formulation with the contravariant base vectors
gi V is given by
(4.4.52)
S = Sji gi gj .
Then the vector w in index notation with the results of the equations (4.4.49), (4.4.51) and
(4.4.52) is rewritten as
w = S (Tu) = Sji gi gj T mk uk gm
j
gi
= Sji T mk uk m
= Sji T jk uk gi = wi gi ,
(4.4.43)
and the coefcients of the vector are given by
(4.4.44)
(4.4.45)
(4.4.53)
For the second order tensor product ST exists in general four representations with all possible
combinations of base vectors
i
T mk gi gk
S T = Sm
ST=
(4.4.46)
and a relation between the transformation tensors B and the retransformation tensor A,
k
A m.
(4.4.49)
Furthermore exists a relation between the transformation tensors A and the retransformation
tensor B
Ami Bmk = ik
(4.4.48)
= T mk ur gkr gm
wi = Sji T jk uk .
BB = 1 , or
m
, and T V V .
Bmi B k = ki
u, v V
= T mk ur (gk gr ) gm ,
(4.4.39)
and the contravariant base vectors are represented by the transformed (overlined) covariant base
vectors,
km
(4.4.40)
gi = Bgi , and B = gk gm gk gm = B gk gm ,
ki
gi = gk gi gk = B gk .
(4.4.41)
, with
In index notation with the covariant base vectors gi V equation (4.4.48) denotes like that
v = T mk gm gk (ur gr )
With this relationship the contravariant base vectors are represented by the transformed (overlined) contravariant base vectors,
m
B k gk
105
ST=
Sim Tkm gi gk
S im Tmk gi gk
covariant basis,
(4.4.54)
contravariant basis,
(4.4.55)
mixed basis,
(4.4.56)
mixed basis.
(4.4.57)
and
(4.4.47)
S T = Sim T mk gi gk
106
Lemma 4.4. The result of the tensor product of two dyads is the scalar product of the inner
vectors of the dyads and the dyadic product of the outer vectors of the dyads. The tensor product
of two dyads of vectors is denoted by
(a b) (c d) = (b c) a d.
(4.4.58)
With this rule the index notation of equations (4.4.53) up to (4.4.57) is easily computed, for
example equation (4.4.53) or (4.4.54) implies
i
ST = Sm
gi gm T nk gn gk
i
T nk nm gi gk ,
= Sm
(4.4.59)
ST 6= TS
(4.4.60)
(ST)T = TT ST
(4.4.69)
(4.4.70)
The linear mappings w = Sv and v = Tu with the vectors u, v, and w V are composed like
w = Sv = S (Tu) = (ST) u = STu = Pu.
(4.4.68)
The "multiplication"or composition of two linear mappings S and T is called a tensor product
P = S T = ST
(4.4.67)
and nally
i
ST = Sm
T mk gi gk .
107
(4.4.71)
(4.4.61)
This "multiplication"is, like in matrix calculus, but not like in "normal"algebra (a b = b a),
noncommutative, i.e.
ST 6= TS.
(4.4.62)
For the three second order tensors R, S, T V V and the scalar quantity R the following
identities for tensor products hold.
(4.4.72)
(4.4.63)
(S) v = (Sv)
(4.4.64)
, with
v V,
(4.4.73)
with this
[ (ST)] v = [S (Tv)] = (S) (Tv) = [(S) T] v,
(4.4.65)
and nally
(ST) = (S) T.
(4.4.66)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.4.74)
108
Proof of equation (4.4.64).
109
, with
(R + S) T = RT + ST,
1 V V,
(R + S) v = Rv + Sv,
(4.4.75)
(4.4.80)
(4.4.76)
and nally
this implies
(R + S) T = RT + ST.
(4.4.81)
(4.4.82)
and nally
1T = T1 = T.
(ST)T = TT ST ,
(4.4.77)
with the dention
, with
which implies
R, S, T V V,
, with
v, w V,
= S,
= ST,
(ST)T
with the assumption of equation (4.4.61) and inserting it into equation (4.4.66),
[(RS) T] v = (RS) (Tv) = (RS) w
ST
(4.4.78)
(4.4.84)
T
(ST)T
= TT ST = ST.
(4.4.83)
(4.4.85)
and nally
(RS) T = R (ST) .
(4.4.79)
(4.4.86)
110
with equations (4.4.64) and (4.4.66),
S1 (ST) (ST)1 = S1 1 = S1 ,
(4.4.87)
S1 (ST) (ST)1 = S1 (ST) (ST)1 ,
(4.4.88)
(4.4.89)
with equation (4.4.88) inserted in (4.4.89) and comparing with equation (4.4.87),
(T) : S = T : (S) = (T : S)
T (ST)1 = S1 ,
T (ST)1 = T1 S1 ,
T : T = tr TTT
= 0 , iff T = 0
T (ST)
(4.4.91)
(4.4.92)
For the norms of tensors, like for the norms of vectors, the following identities hold,
(4.4.90)
= 1 (ST)
(4.4.93)
T : (v w) = vTw
, with
v, w V
, and
T V V .
(a b) : (c d) = (a c) (b d) .
(4.4.94)
For the three second order tensors R, S, T V V and the scalar quantity R the following
identities for scalar products of tensors hold.
S:T=T:S
(4.4.95)
(4.4.96)
(4.4.102)
(4.4.103)
(4.4.104)
With this rule the index notation of equation (4.4.104) implies for example
S : T = S im gi gm : Tnk gn gk
k
= S im Tnk in m
and nally
S : T = S nm Tnm .
(4.4.100)
(4.4.99)
(4.4.101)
|T| = || |T| ,
|T + S| |T| + |S| .
(ST)1 = T1 S1 .
(4.4.98)
= T1 T (ST)
(4.4.97)
S1 (ST) = S1 S T = T,
111
and
(4.4.105)
And for the other combinations of base vectors the results are
S : T = Snm T nm
S : T = S nm Tnm
S : T = Snm T nm .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.4.106)
(4.4.107)
(4.4.108)
112
4.5
It is not absolutely correct to speak about the determinant of a tensor, because it is only the
determinant of the coefcients of the tensor in Cartesian coordinates 1 and not of the whole tensor
itself. For the different notations of a tensor with covariant, contravariant and mixed coefcients
the determinant is given by
det T = det T ij = det [Tij ] = det T ij = det Ti j .
(4.5.1)
Expanding the determinant of the coefcient matrix of a tensor T works just the same as for any
other matrix. For example the determinant could be described with the permutation symbol ,
like in equation (4.2.35)
(4.5.2)
det T = det [Tmn ] = T21 T22 T23 = T1i T2j T3k ijk .
T31 T32 T33
Some important identities are given without a proof by
det (T) = 3 det T,
det (TS) = det T det S,
(4.5.3)
(4.5.4)
det TT = det T,
(det Q) = 1
det T
(4.5.5)
, if Q is an orthogonal tensor,
= (det T)
if T
113
exists.
(4.5.6)
(4.5.7)
The inner product of a tensor T with the identity tensor 1 is called the trace of a tensor,
gk gk : T ij gi gj = T ij ki jk = Tkk ,
(4.5.8)
(4.5.9)
(4.5.10)
1
To compute the determinant of a second order tensor in general coordinates is much more complicated, and this
is not part of this lecture/script, for any details see for example DE B OER [3].
(4.5.11)
tr STT = S : T
(4.5.12)
and easy to proof just writting it in index notation. Starting with this some more important
identities could be found,
tr T = tr TT ,
tr (ST) = tr (TS) ,
tr (RST) = tr (TRS) = tr (STR) ,
tr [T (R + S)] = tr (TR) + tr (TS) ,
tr [(S) T] = tr (ST) ,
(
> 0 , if T 6= 0,
(4.5.13)
(4.5.14)
(4.5.15)
(4.5.16)
(4.5.17)
(4.5.18)
(4.5.19)
|S : T| |S| |T| .
(4.5.20)
Like for the symmetric and skew parts of a tensor there are also a lot of notations for the volumetric and deviator parts of a tensor. The volumetric part of a tensor in the 3-dimensional Euclidean
vector space E3 is dened by
TV = Tvol =
and in this way it is easy to see that the result is a scalar. For the dyadic product of two vectors
the trace is given by the scalar product of the two involved vectors,
tr (a b) = 1 : (a b) = a (1 b) = a b,
gk gk : ai gi bj gj = ai bj ki jk = ak bk .
tr T = 1 : T = T : 1.
1
(tr T) 1
n
, and
T E 3 E3 .
(4.5.21)
It is important to notice that all diagonal components V (i)(i) are equal and all the other components equals zero
(4.5.22)
V ij = 0 if i 6= j.
The deviator part of a tensor is given by
TD = Tdev = dev T = T Tvol = T TV = T
1
(tr T) 1.
n
(4.5.23)
114
115
There are a lot of different notations for the symmetric part of a tensor T, for example
v, w V ,
w (T v) = v (TT w) , and
T V V.
(4.5.24)
TS = Tsym = sym T,
(4.5.41)
w [(a b) v] = v [(b a) w] ,
(4.5.25)
(a b)T = (b a) .
(4.5.26)
TA = Tasym = skew T.
(4.5.42)
and
A second rank tensor is said to be symmetric, if and only if
T = TT .
(4.5.43)
w [(a b) v] = (w a) (b v) ,
(4.5.27)
T = TT .
(4.5.28)
are equal, q.e.d. For the transpose of a tensor the following identities hold,
(a b)T = (b a),
(T ) = T,
(4.5.30)
(4.5.31)
1 = 1,
(S + T)T = ST + TT ,
T
(4.5.32)
(T) = T ,
T
T ij = T ji
(4.5.29)
T T
(4.5.33)
T
(S T) = T S .
(4.5.34)
(4.5.37)
T=
Tij gi
gj T =
Tij gj
g =
Tji gi
g ,
(Tij )T = Tji ,
(4.5.36)
T ij
= Tj i
, or
Ti j
= T ji .
(4.5.45)
(4.5.46)
1
1
T = TS + TA = TTsym + TTasym = sym T + skew T = (T + TT ) + (T TT ). (4.5.47)
2
2
The symmetric part of a tensor is dened by
(4.5.48)
(4.5.38)
(4.5.49)
The inverse of a tensor T exists, if for any two vectors v and w the expression,
and
, if T is antisymmetric.
1
TS = Tsym = sym T = (T + TT ),
2
(4.5.35)
, if T is symmetric,
Any second rank tensor can be written as a sum of a symmetric tensor and an antisymmetric
tensor,
(4.5.44)
(4.5.40)
w = Tv,
(4.5.50)
v = T1 w.
(4.5.51)
could be transformed in
116
117
, and
1 1
= T.
det T 6= 0.
(4.5.52)
T = RU
(4.5.53)
1 1
T ,
(4.5.54)
, or T = VR
, with
TVV
, and
det T 6= 0.
(4.5.62)
(4.5.55)
dz
QQT = QT Q = 1 , i.e.
Q1 = QT .
(4.5.56)
dX
F=T
(4.5.57)
R
The orthogonal mappings of two arbitrary vectors v and w is rewritten with the denition of the
transpose (4.5.24)
(Qw) (Qv) = wQT Qv,
dx
dz
(4.5.58)
(Qw) (Qv) = w v.
(4.5.59)
The scalar product of two vectors equals the scalar product of their orthogonal mappings. And
for the square value of a vector and its orthogonal mapping equation (4.5.59) denotes
(Qv)2 = v2 .
(4.5.60)
Sometimes even equation (4.5.59) and not (4.5.56) is used as denition of an orthogonal tensor.
The orthogonal tensor Q describes a rotation. For the special case of the Cartesian basis the
components of the orthogonal tensor Q are given by the cosine of the rotation angle
Q = qik ei ek
k ))
qik = cos (^ (ei ; e
vectors
and
det Q = 1.
If det Q = +1, then the tensor is called proper orthogonal tensor or a rotator.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.5.61)
dz = R dX,
(4.5.63)
dx = V dz.
(4.5.64)
and
(4.5.65)
118
119
t3
The other possible way to describe the composition is with the vector d z,
dx = R dz,
(4.5.66)
da3
g3 6
g2
and
dz = U dX,
(4.5.67)
dx = R U dX = F dX,
(4.5.68)
and nally
The composed tensor F is called the deformation gradient and its polar decomposition is given
by
T F = R U = V R.
(4.5.69)
g1
T = T ik
gi
gk
|gi | |gk |
T ik
T ik
gi g k ,
=
g(i)(i) g(k)(k)
= T ik g(i)(i) g(k)(k) .
(4.5.71)
(4.5.72)
ik
The stress tensor T = T gi gk is given w.r.t. to the bais gi . Then the associated stress vector
ti w.r.t. a point in the sectional area dai is dened by
ti =
df
da(i)
ti = ik gk
df i = ti da(i) ,
; df i = ik gk da(i) ,
(4.5.74)
with the differential force df . Furthermore the sectional area and its absolute value are given by
(4.5.75)
and
|dai | = da(i) g(i) ,
p
|dai | = da(i) g (i)(i) .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.5.77)
p
p
g (i)(i)
ik
ik
da(i) g (i)(i) = 1,
g(k)(k)
and nally the denition for the physical components of the stress tensor ik is given by
(4.5.73)
gk
da(i) ,
|gk |
p
gk
df i = ik
da(i) g (i)(i) .
g(k)(k)
df i = ik
g(k)(k)
ik = ik p
.
g (i)(i)
(4.5.78)
An isotropic tensor is a tensor, which has in every rotated coordinate system, if it is a Cartesian
coordinate system the same components. Every tensor with order zero, i.e. a scalar quantity, is
an isotropic tensor, but no rst order tensor, i.e. a vector, could be isotropic. The unique isotropic
second order tensor is the Kronecker delta, see section (4.1).
(4.5.76)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
120
4.6
121
(4.6.2)
T = TT = T ki ei ek .
(4.6.3)
The stress vector in the section surface, given by e1 = e1 , is dened by the linear mapping of
the normal unit vector e1 with the tensor T
invariants,
t1 = T (e1 )
(4.6.4)
= T ik (ei ek )e1
= T ik k1 ei ,
t1 = T 1i e1 .
principal axes
(4.6.5)
is described for the problem of the principal axes of a stress tensor, also called the directions of
principal stress. The Cauchy stress tensor in Cartesian coordinates is given by
The stress tensor T assigns the resulting stress vector t(n) to the direction of the normal vector
perpendicular to the section surface n. This linear mapping fullls the equilibrium conditions
T = T ik ei ek .
t(n) = T n,
(4.6.1)
This stress tensor T is symmetric because of the equilibrium conditon of moments. With this
(4.6.6)
ne =
t(n) dA(n)
x3 6
nl lk
(4.6.7)
k
= n = cos n, e
|n| = 1.
t(3) dA(3)
I
t(1) dA(1)
e3 6
t(n) = T ik ei ek nl el
= T ik nl kl ei ,
e2
x2
e1
R
t(2) dA(2)
1
x
(4.6.8)
= T ik nl (ek el ) ei
(n)
= T ik nk ei .
(4.6.9)
The action of the tensor T on the normal vector n reduces the order of the second order tensor
T (stress tensor) to a rst order tensor t (n) (i.e. the stress vector in direction of n).
Lemma 4.6. Principal axes problem Exists a direction no in space, in such a way, that the
resulting stress vector t(n0 ) is oriented in this direction, i.e. that the vector n0 fullls the following
equations?
t(n0 ) = n0
= 1 n0 .
(4.6.10)
(4.6.11)
122
(4.6.12)
T ei ek ik ei ek nl0 el
ik
T ik nl0 (ek el ) ei
ik
T ik nl0 kl ei
ik
T ik n0k ei
= 0,
(4.6.13)
= 0,
(4.6.14)
This is a linear homogenous system of equations for n01 , n02 and n03 , then non-trivial solutuions
exist, if and only if
det(T 1) = 0,
(4.6.17)
or in index notation
(4.6.18)
= 0,
(4.6.19)
= 0,
= 0,
= 0,
Tik g ik = 0,
det Tik g ik = 0.
Tki ki nk0 = 0,
i
det Tk ki = 0.
This equation could be represented in matrix notation, because it is given in a Cartesian basis,
ik
(4.6.15)
T ik [n0k ] = [0] ,
(T 1) n0 = 0,
11
n01
T 12
T 13
0
T
22
23
T 21
T
T
n02 = 0 .
(4.6.16)
T 31
T 32
T 33
n03
0
det(T ik ik ) = 0.
Tik gi gk g ik gi gk nl0 gl
Tik g ik n0k gi
= 0,
T ik ik n0k = 0.
123
(4.6.20)
= 0,
and nally
(4.6.21)
(4.6.22)
For a Cartesian basis the equation (4.6.22) becomes a cubic equation, because of being an eigenvalue problem in E3 with the invariants given by
I1 = tr T = gik Tik = ik T ik = Tkk = T kk ,
1 ii kk
1
(tr T)2 tr (T)2 =
T T T ik T ki ,
I2 =
2
2
I3 = det T = det T ik .
(4.6.23)
I1 = 1 + 2 + 3 ,
I2 = 1 2 + 2 3 + 3 1 ,
I3 = 1 2 3 .
(4.6.26)
(4.6.27)
(4.6.28)
(4.6.24)
(4.6.25)
The fundamental theorem of algebra implies that there exists three roots 1 , 2 , and 3 , such that
the following equations hold
124
125
4.6.7 Example
The assumption that the tensor is symmetric is denoted in tensor and index notation like this
TT = T, T ik = T ki and resp. in matrix notation T T = T . The eigenvalue problem with
j 6= l is given in matrix notation by
Compute the characteristic polynomial, the eigenvalues and the eigenvectors of this matrix
1 4 8
4 7 4 .
8
4 1
(T 1) n0 = 0,
(4.6.29)
+nT0l |
(T j 1) n0j = 0,
(4.6.30)
nT0j |
(T l 1) n0l = 0.
(4.6.31)
(4.6.32)
i.e.
j ) nT0j n0l
0.
(4.6.33)
(4.6.34)
(l j ) nT0j n0l = 0,
bT b + cT c 6= 0,
(4.6.37)
(4.6.38)
i.e. = 0 and the eigenvalues are real numbers. The result is a symmetric stress tensor with
three real principale stresses and the associated directions being orthogonal to each other.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.6.39)
(4.6.35)
(4.6.36)
The coordinates of the associated eigenvectors n0j and n0l could be written as column matrices
n0j and n0l . Furthermore the relations n0j = b + ic and n0l = b + ic hold. Comparing this with
the equation (4.6.34) implies
2i (b + ic)T (b ic) = 0,
2i bT b + cT c = 0,
1 = 9, 2,3 = 9.
This equation holds, if and only if l j 6= 0 and nT0j n0l = 0. The conclusion of this relation
between the eigenvalues j , l , and the normal unit vectors n0j , and n0l is, that the normal vectors
are orthogonal to each other.
j = + i,
l = i.
(T 1) n0 = 0
is substituted by the eigenvalue problem in general coordinates given by
Tik gi gk g ik gi gk nl0 gl = 0,
(4.6.41)
(4.6.42)
with the vector n0 = nl0 gl in the direction of a principal axis. Begining with the eigenvalue
problem
(4.6.43)
Tik g ik n0k gi = 0,
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
126
4.7
and with
n0k 6= 0,
results nally the condition
ik
ik
(4.6.44)
n0k = 0.
(4.6.45)
(4.6.46)
v = Tu
Tki ki nl0 gk gl gi = 0
i
Tk ki nl0 lk gi = 0,
Tki ki n0k = 0.
, with
u, v V , and
T V V.
(4.7.1)
(4.7.2)
T = T jk gj gk
127
, with
gj , gk V.
(4.7.3)
(4.6.48)
= T jk ui gki gj ,
(4.6.49)
But the matrix [Tki ] = T is nonsymmetric. For this reason it is necessary to control the orthogonality of the eigenvectors by a decomposition.
(4.7.4)
= T jk ui (gk gi ) gj
v = Tij ui gj = v j gj .
(4.7.5)
(4.7.6)
A = Ajkl gj gk gl
, with
gj , gk , gl V.
(4.7.7)
(4.7.8)
(4.7.9)
128
E = eijk ei ej ek
, or
E = eijk ei ej ek .
(4.7.10)
ijk
For the orthonormal basis ei the components of the covariant eijk and contravariant e permutation tensor (symbol) are equal
(4.7.11)
ei ej = eijk ek .
Equation (4.2.26) in section 3(4.2) is the short form of a product in Cartesian coordinates between the third order tensor E and the second order identity tensor. This scalar product of the
permutation tensor and ei ej from the right-hand side yields,
ei ej = ersk er es ek : (ei ej )
= ersk is jk er = erij er ,
ei ej = eijr er ,
(4.7.12)
(4.7.13)
= C.
(4.7.18)
Comparing this with the well known unidimensional Hookes law it is easy to see that this mapping is the generalized 3-dimensional linear case of Hookes law. The elasticity tensor C has in
general in space E3 the total number of 34 = 81 components. Because of the symmetry of the
strain tensor and the stress tensor this number reduces to 36. With the potential character
of the elastic stored deformation energy the number of components reduces to 21. For an elastic
and isotropic material there is another reduction to 2 independent constants, e.g. the Youngs
modulus E and the Poissons ratio .
B=
n
X
i=1
ai b i c i =
n
X
i=1
Ti ci = B ijk gi gj gk ,
(4.7.19)
n
X
i=1
ai b i c i d i =
n
X
i=1
Si Ti = C ijkl gi gj gk gl .
(4.7.20)
For example the tensors form order zero till order four are summarized in index notation with a
basis gi ,
Really important is the so called elasticity tensor C used in elasticity theory. This is a fourth
order tensor, which maps the strain tensor onto the stress tensor ,
C=
= ersk ir js ek ,
ei ej = eijk e .
129
and a simple fourth order tensor and a complete fourth order tensor,
ei ej = (ei ej ) : ersk er es ek
k
(0)
= ,
(1)
(4.7.21)
i
v = v = vi g ,
(4.7.22)
(2)
T = T = T jk gj gk ,
(4.7.23)
C = C = C ijkl gi gj gk gl .
(4.7.25)
T=T
mn
(gm gn ) ,
S = C : T = C ijkl gi gj gk gl : (T mn gm gn ) ,
(4.7.15)
B = B ijk gi gj gk ,
(4.7.16)
S = S ij gi gj .
(4.7.17)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(4.7.24)
130
Chapter 5
Vector and Tensor Analysis
S IMMONDS [12], H ALMOS [6], A BRAHAM, M ARSDEN, and R ATIU [1], and M ATTHEWS [11].
And in german DE B OER [3], S TEIN ET AL . [13], and I BEN [7].
131
132
5.1
133
5.2
5.1.2
5.1.3
5.1.4
5.1.5
5.1.6
5.1.7
= ()
()
v=v
(5.1.1)
(5.1.2)
()
T=T
(5.1.3)
and
5.2.2
5.2.3
5.2.4
5.2.5
5.2.6
5.2.7
5.2.8
5.2.9
A scalar function could be represented by another scalar quantity, a vector, or even a tensor,
which depends on one scalar variable. These different types of scalar functions are denoted by
Denitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
5.3.2
5.3.3
5.3.4
5.3.5
5.3.6
The usual derivative w.r.t. a scalar variable of the equation (5.1.1) is established with the
Taylor series of the scalar-valued scalar function () at a value ,
(5.1.4)
( + ) = () + () + O 2 .
The term () given by
() = lim
( + ) ()
,
(5.1.5)
is the derivative of a scalar w.r.t. a scalar quantity. The usual representations of the derivatives
dd are given by
d
( + ) ()
= 0 = lim
.
(5.1.6)
() =
0
d
The Taylor series of the scalar-valued vector function, see equation (5.1.2), at a value is given
by
( + ) = v
() + y () + O 2 .
v
(5.1.7)
The derivative of a vector w.r.t. a scalar quantity is dened by
y () =
( + ) v
()
dv
v
= v0 = lim
.
0
d
(5.1.8)
() is given by
The total differential or also called the exact differential of the vector function v
dv = v0 d.
(5.1.9)
() at a value is given by
The second derivative of the scalar-valued vector function v
d dv
dy ()
=
= v00 .
(5.1.10)
d
d d
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
134
The Taylor series of the scalar-valued tensor function, see equation (5.1.3), at a value is given
by
( + ) = T
() + Y () + O 2 .
(5.1.11)
T
This implies the derivative of a tensor w.r.t. a scalar quantity,
( + ) T
()
dT
T
Y () =
= T0 = lim
.
0
d
(5.1.12)
(5.1.13)
0
(vw) = v w + vw ,
(v w) = v0 w + v w0 ,
0
(v w) = v w + v w ,
(Tv)0 = T0 v + Tv0 ,
0
135
With this partial derivatives ,i the exact differential of the function is given by
d = ,i di .
(ST) = S T + ST ,
1 0
T
= T1 T0 T1 .
(5.1.15)
(5.1.16)
(5.1.17)
(5.1.18)
(5.1.19)
(1 , . . . , i + , . . . , n ) v
(1 , . . . , i , . . . , n )
v
v
,
= v,i = lim
0
i
dv = v,i di .
(1 , . . . , i , . . . , n )
(1 , . . . , i + , . . . , n ) T
T
T
,
= T,i = lim
0
i
dT = T,i di .
(5.1.28)
A vector function x = x (1 ) with one variable 1 in the Euclidean vector space E3 could be
represented by a space curve. The vector x (1 ) is the position vector from the origin O to the
point P on the space curve. The tangent vector t (1 ) at a point P is then dened by
Like for the functions of one scalar varaible it is also possible to denite varoius functions of
more than one scalar variable, e.g.
(5.1.20)
(5.1.21)
dx (1 )
.
t 1 = x 0 1 =
d1
(5.1.29)
The tangent unit vector or just the tangent unit of the space curve at a point P with the position
vector x is dened by
t (s) =
dx
x 1
=
1 s
ds
, and
|t (s)| = 1.
(5.1.30)
The normal vector at a point P on a space curve is dened with the derivative of the tangent
vector w.r.t. the curve parameter s by
and nally
(1 , 2 , . . . , i , . . . , n )
T=T
(5.1.27)
(5.1.26)
The partial derivatives of the tensor-valued function (5.1.22) w.r.t. the scalar variable i are
dened by
= (1 , 2 , . . . , i , . . . , n )
(1 , 2 , . . . , i , . . . , n )
v=v
(5.1.25)
As a short example for a proof of the above identities the proof of equation (5.1.19) is given by
TT1 = 1,
0
0
TT1 = T0 T1 + T T1 = 0,
0
T0 T1 = T T1 ,
0
T1 = T1 T0 T1 .
(5.1.24)
The partial derivatives of the vector-valued function (5.1.21) w.r.t. the scalar variable i are
dened by
(5.1.14)
0
0
(5.1.22)
In stead of establishing the total differentials like in the section before, it is now necessary to
establish the partial derivatives of the functions w.r.t. the various variables. Starting with the
scalar-valued function (5.1.20), the partial derivative w.r.t. the i-th scalar variable i is dened
by
(1 , . . . , i + , . . . , n ) (1 , . . . , i , . . . , n )
.
(5.1.23)
= ,i = lim
0
i
n=
d2 x
dt
= 2.
ds
ds
(5.1.31)
The term 1/ is a measure of the curvature or just the curvature of a space curve at a point P .
The normal vector n at a point P is perpendicular to the tangent vector t at this point,
nt i.e.
n t = 0,
136
137
n M
s = s (1 )
e3 6
x (1 )
R
1
t (1 )
x (1 + 1 )
e1
e2
M
The so called binormal unit vector b or just the binormal unit is the vector perpendicular to the
tangent vector t, and the normal vector n at a point P , and dened by
(5.1.32)
b = t n.
The absolute value |b| of the binormal unit is a measure for the torsion of the curve in space at a
point P , and the derivative of the binormal vector w.r.t. the curve parameter s implies,
The proof of this assumption starts with the scalar product of two tangent vectors,
t t = 1,
d
(t t) = 0,
ds
dt
2 t = 0,
ds
dt
dn
db
=
n+t
ds
ds
ds
dn
=nn+t
ds
1
=0+ n
db
1
= n.
ds
and nally results, that the scalar product of the derivative w.r.t. the curve parameter and the
tangent vector equals zero, i.e. this two vectors are perpendicular to each other,
With the curvature 1 the normal unit vector n or just the normal unit is dened by
3
1
dx d2 x
dx
= 2
2 3.
ds
ds
ds
n = n.
(5.1.35)
dt
t.
ds
1
= n
(5.1.34)
(5.1.33)
(5.1.36)
The three unit vectors t, n and b form the moving trihedron of a space curve in every point P .
The derivatives w.r.t. the curve parameter ds are the so called Serret-Frenet equations, and given
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
138
below,
139
ds
1
1
1
1
dn
= n b t n = t b,
ds
db
1
d
= n = (t b) .
ds
ds
a3 = 0
(5.1.37)
(5.1.38)
a = det a, .
The absolute value of a line element dx is computed by
ds2 = dx x = x, d x, d ,
e3 6
a1
e1
q
a d d .
(5.1.45)
(5.1.46)
The contravariant base vectors of the curved surface are computed with the metric coefcients
and the covariant base vectors,
a = a a ,
(5.1.47)
and nally
dA = a d1 d2 .
P
z
= a a d d = a d d ,
ds =
2
1ds
(5.1.44)
(5.1.39)
a3 o
(5.1.43)
a2
, and a33 = 1,
e2
a a = .
(5.1.48)
x = x i ei .
a1 =
a = a a
, and , = 1, 2.
(5.1.41)
The normal unit vector of the curved surface, perpendicular to the vectors a 1 , and a2 , is dened
by
a1 a 2
n = a3 =
.
(5.1.42)
|a1 a2 |
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(5.1.49)
The curvilinear coordinates or resp. the curvilinear coordinate system is introduced by the
following relations between the curvilinear coordinates i and the Cartesian coordinates xj and
base vectors ej ,
i x1 , x 2 , x 3 .
(5.1.50)
i =
The inverses of this relations in the domain are explicit dened by
xi = xi 1 , 2 , 3 ,
(5.1.51)
140
141
3
g2M
g3
1
2
2
1
x3
3
e3 6
e3 6
x
-
e1
O e2
g1
x
-
x2
e1
O e2
x1
Figure 5.4: Curvilinear coordinates in a Cartesian coordinate system.
x = xi 1 , 2 , 3 ei .
gk gi = ki .
x = xi gi .
(5.1.53)
(5.1.56)
(5.1.57)
The covariant coordinates x and the contravariant coordinates xi of the position vector x are
connected by the metric coefcients like this,
i
A basis in the point P represented by the position vector x and tangential to the curvilinear
coordinates i is introduced by
xi (1 , 2 , 3 )
ei .
k
For each covariant natural basis gk an associated contravariant basis with the contravariant base
vectors of the natural basis g i is dened by
The vector x w.r.t. the contravariant basis is represented by
gk =
position vectors of the points along this curvilinear coordinate. The vector x w.r.t. the covariant
basis is given by
x = xi gi .
(5.1.55)
(5.1.54)
xi = gik xk ,
xi = g ik xk .
(5.1.58)
(5.1.59)
These base vectors gk are the covariant base vectors of the natural basis and form the so called
natural basis. In general these base vectors are not perpendicular to each other. Furthermore
this basis gk changes along the curvilinear coordinates in every point, because it depends on the
The derivative of a covariant base vector gi E3 w.r.t. a coordinate k is again a vector, which
could be described by a linear combination of the base vectors g 1 , g2 , and g3 ,
gi
!
= gi,k = sik gs .
(5.1.60)
k
142
The sik are the components of the Christoffel symbol (i) . The Christoffel symbol could be
described by a second order tensor w.r.t. the basis gi ,
(i) = sij gs gj .
(5.1.61)
With this denition of the Christoffel symbol as a second order tensor a linear mapping of the
base vector gk is given by
s
s j
= ij g gk gs = ij k gs ,
and nally
gi,k =
sik gs .
(5.1.63)
Equation (5.1.63) is again the deniton of the Christoffel symbol, like in equation (5.1.60). With
this relation the components of the Christoffel symbol could be computed, like this
gi,k gs = rik gr gs = rik rs = sik .
(5.1.64)
Like by any other second order tensor, the raising and lowering of the indices of the Christoffel
symbol is possible with the contravariant metric coefcients g ls , and with the covariant metric
coefcients g ls , e.g.
(5.1.65)
ikl = gls sik .
Also important are the relations between the derivatives of the metric coefcients w.r.t. to the
coordinates i and the components of the Christoffel symbol,
ikl =
1
(gkl,i + gil,k gik,l ) .
2
(5.1.66)
5.2
143
(5.2.1)
(5.2.2)
(x)
T=T
(5.2.3)
and
For example some frequently used elds in the Euclidean vector space are,
scalar elds - temperature eld, pressure eld, density eld,
vector elds - velocity eld, acceleration eld,
tensor elds - stress state in a volume element.
(5.2.6)
(x) of a vector-valued vector function (vector eld) is a tensor-valued funcThe gradient grad v
tion (second order tensor depending on the vector of position x). For a scalar-valued vector
function or a scalar eld
(x) there exists an analogue to equation (5.2.4), resp. (5.2.5),
, and |y| 0,
(5.2.7)
(x + y) =
(x) + l (x) y + O y2
(x + y)
(x)
l (x) 1 = lim
.
(5.2.8)
|y|0
|y|
And with this relation the gradient of the scalar eld is a vector-valued vector function (vector
eld) given by
l (x) = grad
(x) .
(5.2.9)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
144
(x) the relations analogue to equaFinally for a tensor-valued vector function or a tensor eld T
tion (5.2.4), resp. (5.2.5), are given by
3
(x + y) = T
(x) + L (x) y + O y
T
(x + y) T
(x)
3
T
.
L (x) 1 = lim
|y|0
|y|
, and
|y| 0,
(5.2.10)
(5.2.11)
The gradient of the second order tensor eld is a third order tensor-valued vector function (third
order tensor eld) given by
3
(x) .
L (x) = grad T
(5.2.12)
The gradient of a second order tensor grad (v w) or grad T is a third order tensor, because of
(grad v) w being a dyadic product of a second order tensor and a vector ("rst order tensor")
is a third order tensor. For arbitrary scalar elds , R, vector elds v, w V, and tensor
eld T V V the following identities hold,
grad ( ) = grad + grad ,
grad (v) = v grad + grad v,
grad (T) = T grad + grad T,
(5.2.13)
(5.2.14)
(5.2.15)
(5.2.16)
(5.2.17)
(5.2.18)
It is important to notice, that the gradient of a vector of position is the identity tensor,
grad x = 1.
(5.2.19)
Most of the up now discussed relations hold for all n-dimensional Euclidean vector spaces E n ,
but most uses are in continuum mechanics and in the 3-dimensional Euclidean vector space E 3 .
The scalar-valued, vector-valued or tensor-valued functions depend on a vector x V, e.g. a
vector of position at a point P . In the sections below the following basis are used, the curvilinear
coordinates i with the covariant base vectors,
x
= x,i ,
i
(5.2.20)
(5.2.23)
(5.2.24)
The denition of the gradient, equation (5.2.1), compared with equation (5.2.23) shows, that the
Christoffel symbols are computed by
(i) = grad gi .
(5.2.25)
Proof. The proof of this relation between the gradient of the base vectors and the Christoffel
symbols,
(5.2.26)
grad gi = (i) = gi,j gj ,
is given by
j
= g gk gi,j = kj gi,j
gi,k = gi,k .
(5.2.27)
145
gi =
(5.2.21)
grad =
i
g = ,i gi ,
i
(5.2.29)
146
than the gradient of the contravariant coefcients v i (x) in the rst term of equation (5.2.28) is
given by
i k
g .
(5.2.30)
grad v i = v,k
Equation (5.2.30) in (5.2.28), and together with equation (5.2.25) the complete gradient of a
vector eld could be given by
i k
grad v = gi v,k
g + v i i ,
(5.2.31)
i
v,k
gi
g +
v i sik gs
g .
(5.2.32)
147
This denition implies another notation for the gradient in a 3-dimensional Cartesian basis of the
Euclidean vector space,
grad = .
(5.2.39)
, with
vi = vi (x1 , x2 , x3 ) ,
(5.2.40)
than the gradient of the vector eld in a Cartesian basis is given with the relation of equation
(5.2.14) by
(x) = grad (vi ei ) = ei grad vi + vi grad ei .
grad v = grad v
(5.2.41)
Computing the second term of equation (5.2.41) implies, that all derivatives of base vectors w.r.t.
the vector of position x, see the denition (5.2.38), are equal to zero,
i s , and s i.
Rewritting equation (5.2.32), and factor out the dyadic product, implies
i
+ v s isk gi gk .
grad v = v,k
(5.2.33)
(ei )
(ei )
(ei )
( )
e1 +
e2 +
e3 =
ei = 0.
x1
x2
x3
xi
(5.2.42)
The term
i
+ v s isk ,
v i |k = v,k
(5.2.34)
grad v = ei grad vi + 0 = ei vi,k ek ,
(5.2.43)
is called the covariant derivative of the coefcient v i w.r.t. the coordinate k and the basis gi .
Than the gradient of a vector eld is given by
i
, with v i |k = v,k
+ v s isk .
(5.2.35)
grad v = v i |k gi gk
and nally the gradient of a vector eld in a 3-dimensional Cartesian basis of the Euclidean
vector space is given by
, with
x = x i ei = x i e i = x i e i ,
(5.2.36)
(5.2.44)
, and
,i =
.
xi
and must be a scalar quantity, because the gradient of a vector is a second order tensor, and the
trace of a second order denes a scalar quantity. The divergence of a tensor eld is dened by
(. . .)
(. . .)
(. . .)
e1 +
e2 +
e3 ,
x1
x2
x3
( )
ei .
xi
(5.2.45)
(5.2.37)
(5.2.38)
, and T V V
(5.2.46)
and must be a vector-valued quantity, because the scalar product of the second order unit tensor
1 and the third order tensor grad (T) is a vector-valued quantity . Another possible denition is
given by
(5.2.47)
a div T = div TT a = grad TT a : 1.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
148
For an arbitrary scalar eld R, and arbitrary vector elds v, w V, and arbitrary tensor
elds S, T V V the following identities hold,
div (v) = v grad + div v,
div (T) = T grad + div T,
(5.2.48)
(5.2.49)
(5.2.50)
(5.2.51)
(5.2.52)
(5.2.53)
(5.2.54)
(5.2.55)
|k sr gir g ks
div v = v i |i .
= v |k gis g
ks
=v
|k ik ,
(5.2.57)
and
(5.2.59)
1 = sr gr gs .
(5.2.62)
(5.2.63)
is the so called covariant derivative of the tensor coefcients w.r.t. the coordinates l , than the
divergence is given by
(5.2.64)
div T = T kl |l gk .
Other representations are possible, e.g. a mixed formulation is given by
div T = T kl |k gl ,
(5.2.65)
l
T kl |k = Tl,k
Tnk nlk + Tln kkn
(5.2.66)
T kl |k ,
A vector eld v in a Cartesian basis e i E3 is represented by equation (5.2.40) and its gradient
by equation (5.2.41). The divergence dened by (5.2.45) rewritten with using the denition of
the gradient of a vector eld in a Cartesian basis (5.2.44) is given by
T ik = Tik 1 , 2 , 3 , . . . , n ,
(5.2.61)
(x) = T ik (gi gk ) V V,
T=T
149
The divergence of a second order tensor is a vector, see also equation (5.2.52),
div T = div T ik gi gk
ik
= grad T gi gk + [div gk ] T ik gi ,
(5.2.60)
div v =
v1
v2
v3
+
+
.
x1 x2 x3
(5.2.68)
150
(5.2.69)
In English textbooks in most cases the curl operator curl instead of the rotation operator rot is
used,
rot v = curl v.
(5.2.70)
The rotation, resp. curl, of a vector eld rot v (x) or curl v (x) is a unique vector eld. Sometimes another denition of the rotation of a vector eld is given by
rot v = div (1 v) = 1 grad v.
(5.2.71)
For an arbitrary scalar eld R, and arbitrary vector elds v, w V, and an arbitrary tensor
eld T V V the following identities hold,
rot grad = 0,
div rot v = 0,
rot grad v = 0,
(5.2.72)
(5.2.73)
(5.2.74)
div (rot T) = 0,
and T is a tensor-valued quantity. For an arbitrary vector eld v V, and an arbitrary tensor
eld T V V the following identities hold,
h
i
div grad v (grad v)T = v grad div v,
(5.2.87)
rot rot v = grad div v v,
tr T = tr T
(5.2.75)
(5.2.76)
(5.2.77)
(5.2.78)
(5.2.80)
(5.2.81)
(5.2.82)
Also important to notice is that, if the tensor eld T is symmetric, then the following identity
holds,
rot T : 1 = 0.
(5.2.83)
(5.2.84)
The laplacian of a scalar eld is a scalar quantity. The laplacian of a vector eld v is
dened by
v = (grad grad v) 1,
(5.2.85)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(5.2.88)
(5.2.89)
(5.2.90)
Finally, if the tensor eld T is symmetric and dened by T = S 1 tr S, with the symmetric
part given by S, then the following identity holds,
(5.2.79)
151
and v is a vector-valued quantity. The denition of the laplacian of a tensor eld T is given
by
T = (grad grad T) 1,
(5.2.86)
(5.2.91)
152
5.3
Integral Theorems
153
resp.
dV =
5.3.1 Denitions
The surface integral of a tensor product of a vector eld u (x) and an area vector da should
be transformed into a volume integral. The volume element dV with the surface dA is given at
a point P by the position vector x in the Euclidean space E3 . The surface dA of the volume
element dV is described by the six surface elements represented by the aera vectors da 1 , . . .,
and da6 .
gg3 g3 d1 d2 d3 =
gd1 d2 d3 .
(5.3.4)
i be dened as the mean value of the vector eld u in the area element i, for example for
Let u
i = 1 a Taylor series is derived like this,
1 = u +
u
u d2
u d3
+
.
2
2
3 2
(5.3.5)
1 +
4 = u
u
dA
6
5
1
u
dV
da1
1
)
4 =
u
1
u
1
1 +
u
1 d
4 = u
1 +
u
d2 g2
3
d g3
1 +
4 = u
u
u
d1 ,
1
u
2 +
5 = u
d2 ,
u
2
4 = u
1 +
u
e1
6 = u
3 +
u
(5.3.7)
(5.3.8)
(5.3.1)
(5.3.2)
and
1
da3 = d d g2 g1 = d d
gg = da6 .
(5.3.3)
The volume of the volume element dV is given by the scalar triple product of the tangential
vectors gi associated to the curvilinear coordinates i at the point P ,
dV = (g1 g2 ) g3 d1 d2 d3 ,
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
u
d3 .
3
(5.3.9)
The surface integral is approximated by the sum of the six area elements,
(5.3.6)
and
O e2
d1 ,
d1 g1
u
2 u d2 1
2 u d3 1
d1 +
d +
d .
1
1
2
2
1 3 2
e3 6
u d3
3 2
and nally
1
3
1 1
u
d ,
1
h
u d2
u +
+
2 2
u da =
dA
6
X
i=1
i dai .
u
(5.3.10)
1 da1 + u
2 da2 + u
3 da3 + u
4 da4 + u
5 da5 + u
6 da6 ,
i dai = u
u
154
and nally with equations (5.3.7)-(5.3.9),
6
X
i=1
i dai =
u
u
u
u
d1 da1
d2 da2
d3 da3 .
1
2
3
(5.3.11)
i dai =
u
u
u
u
1
2
3
g
+
g
+
g
gd1 d2 d3 ,
1
2
3
i=1
volumes are summed over, then the dyadic products of the inner surfaces vanish, because every
dyadic product appears twice. Every area vector da appears once with the normal direction n
and once with the opposite direction n. The vector eld u is by denition continuous, i.e. for
each of the two sides of an inner surface the value of the vector eld is equal. In order to solve
the whole problem it is only necessary to sum (to integrate) the whole outer surface dA with the
normal unit vector n. If the summation over all subvolumes dV with the surfaces da is rewritten
as an integral, like in equation (5.3.13), then for the whole volume and surface the following
relation holds,
Z
Z
nV Z
X
u da = u da = grad u dV ,
(5.3.14)
V
nV Z
X
u
i
dV ,
g
i
i=1
u n da =
dA
u n da =
grad u dV .
(5.3.15)
With this integral theorems it would be easy to develop integral theorems for scalar elds, vector
elds and tensor elds.
155
i=1 dA
i dai = grad u dV .
u
(5.3.12)
dA
u da =
6
X
i=1
i dai =
u
u
gi dV = grad u dV .
i
(5.3.13)
Equation (5.3.13) holds for every subvolume dV with the surface dA. If the terms of the sub-
u n da =
div u dV .
(5.3.16)
Proof. Equation (5.3.15) is multiplied scalar with the unit tensor 1 from the left-hand side,
Z
Z
1 : u n da = 1 : grad u dV ,
(5.3.17)
V
1 : u n = g j g j : uk g k n i g i
= uk ni gjk ij = uj nj ,
1 : u n = u n,
(5.3.18)
and the scalar product of the unit tensor and the gradient of the vector eld
1 : grad u = tr (grad u) ,
1 : grad u = div u,
Figure 5.7: The Volume, the surface and the subvolumes of a body.
(5.3.19)
(5.3.20)
156
(5.3.21)
(5.3.22)
ZA
div TT a dV ,
T n da =
a n da
A
n da =
grad dV = 0,
grad dV .
n u da =
rot u dV ..
(5.3.26)
The Stokes theorem for the cross product of a vector eld u and its normal vector n is dened
by
Z
Z
div T dV .
(5.3.25)
(5.3.24)
and nally
Z
Z
a T n da div T dV = 0,
A
With TT a = u being a vector it is possible to use Gausss theorem (5.3.16), this implies
Z
Z
T
T a n da
a T n da =
A
157
Using the identity (5.2.48) and the vector a being constant yields
(5.3.23)
with the cross product u a being a vector it is possible to use the Gausss theorem (5.3.16)
Z
Z
(5.3.27)
a n u da = div (u a) da.
A
(5.3.28)
158
and nally
Z
n u da =
Chapter 6
rot u dV .
Exercises
i be the mean value in a subvolume Vi . The volume integral of the divergence of the
Let u
vector eld u with inserting the relation of equation (5.3.29) is given by
Z
X
^
div u dV = lim
(5.3.30)
div
ui Vi .
Vi 0
u da
a
.
div u = lim
V 0
V
(5.3.31)
ai
^
div ui =
,
Vi R
u da
Z
X ai i
div u dV = lim
.
Vi 0
Vi
i
V
and nally with the suammation of all subvolumes, like at the begining of this section,
Z
Z
Z
div u dV = u da = u nda.
(5.3.32)
V
159
Chapter 6. Exercises
160
6.1.2
6.1.3
6.1.4
6.1.5
6.1.6
6.1.7
6.1.8
6.2.2
6.5
6.2.3
6.2.4
6.2.5
6.2.6
Transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180
6.7
6.6.2
6.6.3
6.6.4
6.6.5
6.6.6
6.7.2
6.7.3
6.7.4
6.7.5
Invariants . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216
6.7.6
6.7.7
6.7.8
6.4
6.3
Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 196
6.2
6.6
6.3.2
6.3.3
6.3.4
6.3.5
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187
6.3.6
Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188
6.4.2
6.4.3
6.4.4
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 191
6.4.5
Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192
6.5.2
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195
161
162
Chapter 6. Exercises
6.1
S1 = S 2
A very simple truss formed by two bars is given like in gure (6.1), and loaded by an arbitrary
force F in negative y-direction. The discrete values of the various quantities, like the Youngs
, and S1 + S2 =
bar II, l, A2 , E2
y
= 45o
S2
F1x
S1
S1
y
6
S2 R
1
F3x
mdoulus Ei or the sectional area Ai , are of no further interest at the moment. Only the forces in
direction of the bars are to be computed. The equilibrium conditions of forces at the node 2 are
I
(6.1.4)
6F1y
F
.
2 cos
6F3y
(6.1.3)
The equilibrium conditions of forces at the node 1 are given in horizontal direction by
bar I, l, A1 , E1
F
,
cos
and nally
S1 = S 2 =
163
FH = 0 = F1x S1 cos ,
(6.1.5)
FV = 0 = F1y S1 sin ,
(6.1.6)
see also the right-hand side of the free-body diagram (6.3). The rst one of this two relations
yields with equation (6.1.4)
F1x = S1 cos =
F cos
2 cos
, and nally
1
F 1x = F ,
2
(6.1.7)
F sin
2 cos
, and nally
1
F 1y = F .
2
(6.1.8)
(6.1.1)
FV = 0 = F + S2 cos + S1 cos ,
(6.1.2)
F1y = S1 sin =
The equilibrium conditions of forces at the node 3 are given in horizontal direction by
X
FH = 0 = F3x S2 cos ,
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(6.1.9)
Chapter 6. Exercises
164
165
FV = 0 = F3y S2 sin ,
6Fy = 2kN
(6.1.10)
see also the left-hand side of the free-body diagram (6.3). The rst one of this two relations
yields with equation (6.1.4)
F3x = S2 cos =
F cos
2 cos
1
, and nally F 3x = F ,
2
F2y
F sin
2 cos
, and nally
1
F 3y = F .
2
(6.1.12)
Fy = 2kN
4
1 = 30
Fx = 10kN
I
III
h = 5, 0m
II
x1
x2
y
6
1
-
3 = 90
h
x3
Fx = 10kN
SIII
SII
F2x
-
The stresses in normal direction of the bars and the nodal displacements could be computed
with the relations described in one of the following sections, see also equations (6.1.17)-(6.1.19).
The important thing to notice here is, that there are overall 6 equations to solve, in order to
compute 6 unknown quantities F1x , F1y , F3x , F3y , S1 , and S2 . This is characteristic for a statically
determinate truss, resp. system, and the other case is discussed in the following section.
2 = 60
3
SI
SII
x2
(6.1.11)
x
Figure 6.5: Free-body diagrams for the nodes 2 and 4.
nodes! In order to get enough equations for computing all unknown quantities, it is necessary
to use additional equations, like the ones given in equations (6.1.17)-(6.1.19). For example the
equilibrium condition of horizontal forces at node 4 is given by
X
FH = 0 = Fx SII cos 2 SI cos 1 ,
(6.1.13)
and in vertical direction by
X
FV = 0 = Fy SIII SII sin 2 SI sin 1 .
(6.1.14)
The moment equilibrium condition at this point is of no use, because all lines of action cross the
node 4 itself! The equilibrium conditions for every support contain for every node 1-3 two unknown reactive forces, one horizontal and one vertical, and one also unknown force in direction
of the bar, e.g. for the node 2,
X
FH = 0 = F2x + SII cos 2 ,
(6.1.15)
and
(6.1.16)
Finally summarizing all possible and useful equations and all unknown quantities implies, that
there are overall 9 unknown quantities but only 8 equations! This result implies, that it is necessary to take another way to solve this problem, than using the equilibrium conditions of forces!
An arbitrary bar with its local coordinate system x, y is given in gure (6.6), with the nodal forces
fix , and fiy at a point i with i = 1, 2 in the local coordiante system. The following relations hold
Chapter 6. Exercises
166
I f2y
S = Ax =
f2x
(6.1.23)
EA T
f =
q q u,
l
bar, with A, E, l
I f1y
EA
ux ,
l
I y
167
(6.1.24)
f1x
u , with
f = K
0 1 0
0 0 0
.
0 1 0
0 0 0
(6.1.25)
An arbitrary bar with its local coordinate system x, y and a global coordinate system x, y is given
in gure (6.7). At a point i with i = 1, 2 the nodal forces f ix , fiy , and the nodal displacements uix ,
uiy are dened in the local coordiante system. It is also possible to dene the nodal forces f ix , fiy ,
and the nodal displacements vix , viy in the global coordiante system. In order to combine more
in the local coordinate system x, y of an arbitrary single bar, see gure (6.6),
S
,
A
ux
dux
kinematics x =
=
,
d
x
l
material law x = Ex .
= EA 0
K
l 1
0
stresses x =
(6.1.17)
(6.1.18)
(6.1.19)
I y
In order to consider the additional relations given above, it is useful to combine the nodal displacements and the nodal forces for the two nodes of the bar in the local coordinate system. The
nodal displacements in the local coordinate system x are given by
ux = q T u = u1x + u2x ,
I f1y
6v1y
(6.1.20)
and the nodal forces are given by the equilibrium conditions of forces at the nodes of the bar,
f = q S,
(6.1.21)
6f1y
I u1y
y, vy
6
v1x
f1x
f1x
u1x
- x, v
x
with
1
0
q=
1
0
u1x
u1y
, u=
u2x
u2y
, and
f1x
f1y
f =
.
f2x
f2y
Chapter 6. Exercises
168
one global coordinate system. This transformation of the local vector quantities, like the nodal
load vector f , or the nodal displacement vector u, is given by a multiplication with the so called
a
= Q a , with
cos sin
Q=
.
sin cos
(6.1.26)
Because it is necessary to transform the local quantities for more than one node, the transforma
v4x
f = Q f
(6.1.27)
v1x
v1y
, with v =
v2x ,
v2y
f1x
f1y
, with f =
.
f2x
f2y
(6.1.28)
v4y .
(6.1.34)
result is, that it is sufcient to consider only one submatrix K i for every element I-III. For
example the complete equilibrium conditions for bar III are given by
0
f3x
"
# v3x
. . . . . . v3y
0
f3y
= .
=
(6.1.35)
v4x
f4x
. . . K 3 v4x
f4y III
v4y III
v4y III
Finally the equilibrium conditions at node 4 with summarizing the bars i = I-III is given by
X
3 i
3 h
i v
X
f4x
F
4x
i
= x =
,
(6.1.36)
i
K
f4y
Fy
v4y
i=1
(6.1.29)
i=1
Inserting this relations in equation (6.1.25) and multiplying with the inverse of the transformation
matrix from the left-hand side yields in the global coordiante system,
Q v = K v,
f = Q1 K
(6.1.30)
with the symmetric local stiffness matrix given in the global coordinate system by
, and
This implies, that it is sufcient to determine the equilibrium conditions only at node 4. The
The local quantities f , and u in equation (6.1.25) are replaced by the following expressions,
u=Qv
(6.1.32)
(6.1.33)
With this conditions the rigid body movement is eliminated from the system of equations, resp.
the assembled global stiffness matrix, than only the unknown displacements at node 4 remain,
cos sin
0
0
Q 0
sin cos
0
0
.
Q=
=
0
0
cos
sin
0 Q
0
0
sin cos
169
sin cos
cos2
sin cos
cos2
2
EA sin cos
sin
sin cos
sin2
.
K=
sin cos
cos2
sin cos
l cos2
sin cos
sin2
sin cos
sin2
(6.1.31)
(6.1.37)
(6.1.38)
I
II
III
= v4y
= v4y
.
v4y
(6.1.39)
By using this way of assembling the reduced global stiffness matrix the boundary conditions are
already implemented in every element stiffness matrix. The second way to assemble the reduced
global stiffness matrix starts with the unreduced global stiffness matrix given by
i
f1x
v1x
0
i
f1y
i
v1y 0
f2x
0 K 14
K 11 0
v2x 0
3 i
X
f2y 0 K 22 0 K 24 v2y 0
= ,
i =
(6.1.40)
f 0
0 K 33 K 34
v3x 0
i=1 3x
f i
K 41 K 42 K 43 K v3y 0
3y
i
v4x Fx
f4x
i
v4y
Fy
f4y
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
170
K v = P.
(6.1.41)
Each element resp. each bar could be described by a combination i, j of the numbers of nodes
used in this special element. For example for the bar (element) III the submatrices K 33 , K 34 ,
K 43 , and a part of K, like in equation (6.1.36) are implemented in the unreduced global stiffness
matrix. After inserting the submatrices K ij for the various elements and considering the boundary conditions given by equations (6.1.32)-(6.1.33) the reduced global stiffness matrix is given
by
P = K v,
(6.1.42)
resp.
Fx
Fy
EA 3 + 3 1 1 v4x
=
,
1 3 v4y
8h
(6.1.43)
see also equation (6.1.36) and (6.1.37) of the rst way. But this computed reduced global stiffness matrix is not the desired result, because the nodal displacements and forces are the desired
quantities.
P.
(6.1.44)
171
EA 3 + 3
1
(8h)2
3 1
,
K=
2
1 1
det K
8h
2 3 + 3 (EA)2
4h
3 1
K 1 =
.
EA 3 + 3 1 1
(6.1.45)
f i = K ivi.
0 0
EA
3
0 1
K =
h 0 0
0 1
(6.1.46)
(6.1.47)
(6.1.49)
0 0
0 1
0 0
0 1
v3 =
v4x ,
v4y
(6.1.50)
with sin 90 = 1, and cos 90 = 0 in equation (6.1.31). The forces in the bars are given in the
global coordinate system, see equation (6.1.30), by
0
0
f3x
8
4
= 6, 762 = f3y ,
f 3 = K 3v3 =
(6.1.51)
0
f4x
3+ 3 0
8 III
f4y III
6, 762 III
and in the local coordinate system associated to the bar III, see equation (6.1.29), by
f3x
f3y
6, 762
3
f3y f3x 0
f = Q3 f 3 =
f4x = f4y = 6, 762 .
f4y
f4x
0
(6.1.52)
Comparing this result with the relation (6.1.21) implies nally the force S III in direction of the
bar,
(6.1.53)
SIII = f3x = f4x = 6, 762kN ,
and for the bars I and II,
SI = 8, 56kN
and by inserting equations (6.1.46), and (6.1.47) in relation (6.1.44) the nodal displacements at
node 4 are given by
4h
v
28
.
(6.1.48)
v = 4x =
v4y
EA 3 + 3 8
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
The forces in the various bars are computed by solving the relations given by the equation
(6.1.28)-(6.1.31) for each element i, resp. for each bar,
, and
SII = 5, 17kN .
(6.1.54)
Comparing this results as a probe with the equilibirum conditions given by the equations (6.1.13)(6.1.14), in horizontal direction,
X
FH = 0 = Fx SII cos 2 SI cos 1
3
1
+ 6, 762 1 = 4, 7 103 0,
= 2, 0 8, 56 5, 17
(6.1.55)
2
2
and in vertical direction,
X
FV = 0 = Fy SIII SII sin 2 SI sin 1
3
1
5, 17 + 6, 762 0 = 1, 8 103 0.
= 10, 0 8, 56
2
2
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(6.1.56)
Chapter 6. Exercises
172
x+
N x d
pux d
x = 0.
It is easy to see, that the left-hand side is very similar to the stiffness matrices, and that the righthand side is very similar to the load vectors in equations (6.1.36), and (6.1.37), or (6.1.42), and
(6.1.43). This simple example shows the close relations between the principle of virtual works,
the nite element methods, and the matrix calculus described above.
(6.1.57)
The force in normal direction of a bar, see also equations (6.1.17)-(6.1.19), is given by
N = EAx .
With this relation the equation (6.1.57) is rewritten,
Z
Z
x + pux d
x = 0.
W = xT EAx d
(6.1.58)
(6.1.59)
The vectors given by the equations (6.1.22), these are the various quantities w.r.t. the local
variable x in equation (6.1.57) could be described by
displacement
strain
virtual displacement
virtual strain
,
ux = qT u
(6.1.60)
,
u,x = x = qT,x u
(6.1.61)
,
ux = qT u
(6.1.62)
,
u,x = x = qT,x u
(6.1.63)
+
= 0.
=
uT
q,x EAqT,x d
x u
x u
(6.1.65)
pqT d
These integrals just describe one element, resp. one bar, but if a summation over more elements
is introduced like this,
Z
3
3
X
X
i T
i +
i = 0,
q,x EAqT,x d
u
(6.1.66)
pi qT d
W i =
x u
x u
i=1
i=1
3
3 Z
X
i T
i =
i .
q,x EAqT,x d
pi qT d
u
x u
x u
i=1
(6.1.67)
i=1
173
Chapter 6. Exercises
174
6.2
1
k2 = k
2
k = 10
kN
cm
, and l = 200cm.
y
6
FAy
k1
175
3
l
2
C0
SI z
6x1
?
D
SIII
Fcritical
x2
?3
2
FBy
k2
Figure 6.9: The free-body diagrams of the subsystems left of node C, and right of node D after
the excursion.
EJ =
EJ =
x1
3
l
2
Fcritical
x2
EJ =
2l
3
l
2
x
Figure 6.8: The given structure of rigid bars.
Rewrite the system of equations so that the general eigenvalue problem for the critical load
Fc is given by
A x = Fc B x.
Transform the general eigenvalue problem into a special eigenvalue problem.
Calculate the eigenvalues, i.e. the critical loads Fc , and the associated eigenvectors.
Check if the eigenvectors are orthogonal to each other.
Transform the equation system in such a way, that it is possible to compute the Rayleigh
quotient. What quantity could be estimated with the Rayleigh quotient?
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
3
MD = 0 = FBy l + Fc x2
2
3
MC = 0 = FAy l + FAx x1
2
FBy =
2Fc
x2 ,
3l
(6.2.1)
and the moment equation w.r.t. the node C of the subsystem on the left-hand side of node C after
the excursion implies with the following relation (6.2.3),
FAy =
2FAx
2Fc
x1 =
x1 .
3l
3l
(6.2.2)
At any time, and any possible excursion, or for any possible load F c the complete system, cf.
(6.10), must satisfy the equilibrium conditions. The equilibrium condition of forces in horizontal
direction, cf. (6.10), after the excursion is given by
X
FH = 0 = FAx Fc FAx = Fc .
(6.2.3)
The moment equation w.r.t. the node A for the complete system implies
X
7
3
MA = 0 = FBy 5l + k2 x2 l + k1 x1 l,
2
2
0=
2Fc
7
3
x2 5l + kx2 l + kx1 l,
3l
4
2
(6.2.4)
Chapter 6. Exercises
176
6k1 x1
6
6
?x1
FAx 6
C
y
6
FAy
3
l
2
In order to solve this general eigenvalue problem with the aid of the characteristic equation it is
necessary to transform the general eigenvalue problem into a special eigenvalue problem. Thus
the equation (6.2.9) is multiplied with the inverse of matrix B from the left-hand side,
B
6
Fcritical
x2
D
2l
177
k 2 x2
A
-
B 1 |
FBy
?3 l
2
-
A x = Fc B x,
(6.2.10)
B 1 A x = Fc B 1 B x,
after this multiplication both terms are rewritten on one side and the vector x is factored out,
0 = B 1 A x Fc 1 x,
Figure 6.10: The free-body diagram of the complete structure after the excursion.
0 = (C x 1 Fc ) x , with
and nally
(6.2.11)
(6.2.5)
(6.2.6)
and nally
1
2 Fc
(x1 + x2 ) .
kx1 + kx2 =
2
3 l
(6.2.7)
The relations (6.2.5), and (6.2.7) are combined in a system of equations, given by
7
k
4
1
k
2
x1
0
= Fc 2
x2
3l
(6.2.12)
B 1 B = 1
, resp.
a b
c d
0
2
3l
10
3l
2
3l
1 0
.
0 1
3
b = l,
2
d = 0,
2Fc
2Fc
1
x1
x2 + kx2 + kx1 ,
0=
3l
3l
2
k
k
a b
,
c d
(6.2.13)
FV = 0 = FAy + FBy + k2 x2 + k1 x1 ,
B 1 =
10
3l
2
3l
x1
,
x2
(6.2.8)
or in matrix notation
A x = Fc B x.
(6.2.9)
This equation system is a general eigenvalue problem, with the F ci being the eigenvalues, and
the eigenvectors xi0 .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
2
1
3
10
a + b = 0 a = b a = l,
3l
3l
5
10
10
2
3
c + d = 1 c = l,
3l
3l
10
and nally the inverse B 1 is given by
3
l 3l
B 1 = 310 2 .
l 0
10
(6.2.14)
(6.2.15)
(6.2.16)
(6.2.17)
(6.2.18)
The matrix C for the special eigenvalue problem is computed by the multiplication of the two
2 2-matrices B 1 , and A like this,
3
21
9
kl 40
kl
l 3 l 32 k 74 k
C = B 1 A = 310 2
= 20
,
1
9
21
k 2k
l 0
kl 40 kl
10
20
and nally the matrix C for the special eigenvalue problem is given by
3
14 3
C = kl
.
6 7
40
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(6.2.19)
Chapter 6. Exercises
178
179
It is possible to choose for each eigenvalue Fci the rst component of of the associated eigenvector like this,
(6.2.27)
xi01 = 1,
and after this to compute the second components of the eigenvectors,
det (C Fc 1) = 0,
(6.2.20)
21
20
kl Fc
9
kl
20
9
kl
40
21
kl
40
Fc
21
21
kl Fc
kl Fc
20
40
21
441 2 2 21
k l klFc klFc + Fc2
800
20
40
= 0.
(6.2.21)
, resp.
xi02 =
C11 Fci
,
C12
(6.2.28)
xi02 =
C21
xi
C22 Fci 01
, resp.
xi02 =
C21
.
C22 Fci
(6.2.29)
Inserting the rst eigenvalue F c1 in equation (6.2.28) implies the second component x102 of the
rst eigenvector,
81 2 2
k l = 0,
800
81 2 2
k l = 0,
800
21
x102 = 20
kl 65 kl
9
kl
40
2
x102 = ,
3
(6.2.30)
and for the second eigenvalue Fc2 the second component x202 of the second eigenvector is given
by
360 2 2 63
k l klFc + Fc2 = 0.
800
40
k l ,
80
40
4
600
r
1089
63
Fc1/2 = kl
kl,
80
6400
63
33
Fc1/2 = kl kl,
80
80
21
(6.2.22)
x202 = 20
, and
3
Fc2 = kl = 750kN .
8
(6.2.23)
(6.2.24)
(6.2.25)
xi01
0
.
=
0
xi02
x202 = 3.
(6.2.31)
(6.2.32)
The eigenvectors x10 , x20 are computed by inserting the eigenvalues Fc1 , and Fc2 in equation
(6.2.11), given by
(C Fci 1) xi0 = 0,
kl 38 kl
9
kl
40
It is sufcient to compute the scalar product of two arbitrary vectors, in order to check, if this
two vectors are orthogonal to each other, i.e.
C11 Fci i
x01
C12
or
6
Fc1 = kl = 2400kN
5
xi02 =
(6.2.26)
x1 x2
, resp.
x1 x2 = 0.
In this special case the scalar product of the two eigenvectors is given by
1
1
x10 x20 = 2 x20 =
= 1 2 = 1 6= 0,
3
3
(6.2.33)
(6.2.34)
i.e. the eigenvectors are not orthogonal to each other. The eigenvectors for different eigenvalues
are only orthogonal, if the matrix C of the special eigenvalue problem is symmetric 2 . If the
matrix of a special eigenvalue problem is symmetric all eigenvalues are real, and all eigenvectors
are orthogonal. In this case all eigenvalues Fc1 , Fc2 are real, but the matrix C is not symmetric,
and for that reason the eigenvectors x10 , x20 are not orthogonal.
2
Chapter 6. Exercises
180
6.2.6 Transformation
The special eigenvalue problem (6.2.11) includes an anti-symmetric matrix C. In order to determine the Rayleigh quotient it is necessary to have a symmetric matrix in the special eigenvalue
problem 3 ,
(C Fc 1) x = 0,
(6.2.35)
and in detail
21
9
kl
40
21
kl
40
kl
20
9
kl
20
Fc
1 0
0 1
x1
0
=
.
x2
0
(6.2.36)
The rst step is to transform the matrix C into a symmetric matrix. Because the matrices are
such simple it is easy to see, that if the second column of the matrix is multiplied with 2, the
matrix becomes symmetric,
21
9
1 0
x1
0
kl 20
kl
20
F
=
,
(6.2.37)
c
1
9
kl 21
kl
x
0 2
0
20
20
2 2
and in matrix notation with the new dened matrices D, and E 2 , and a new vector q
D Fc E 2 q = 0.
(6.2.38)
Because the matrix E 2 = E T E is a diagonal and symmetric matrix, the matrices E, and E T are
diagonal and symmetric matrices, too,
1 0
1 0
.
(6.2.39)
E2 =
E = ET =
0 2
2
0
1
E D E 1 Fc E 1 E T E E 1 E q = 0,
1
0
1
2
, and E 1 = E 1 .
(6.2.41)
The equation (6.2.38) is again a general eigenvalue problem, but now with a symmetric matrix
D. But in order to compute the Rayleigh quotient, it is necessary to set up a special eigenvalue
problem again. In the next step the identity 1 = E 1 E is inserted in equation (6.2.38), like this,
D Fc E 2 E 1 E q = 0,
(6.2.42)
3
(6.2.43)
E D E 1 Fc 1 1 E q = 0,
(6.2.44)
With relation (6.2.41) the rst term in equation (6.2.44) describes a congruence transformation,
so that the matrix F keeps symmetric 4 ,
T
F = E 1 D E 1 = E 1 D E 1 ,
(6.2.45)
and with the matrix F given by
21
1
0
kl
20
F =
9
1
kl
0 2 2 20
9
kl
20
21
kl
20
1
0
1 0
x1
p = Eq =
1
0
2 2 x2
" 21
kl
0
= 20
9
1
kl
2
2
20 2
p=
9
kl
20 2
21
kl.
40
x1
.
1
2x2
2
(6.2.46)
(6.2.47)
Finally combining this results implies a special eigenvalue problem with a symmetric matrix F
and a vector p,
(F Fc 1) p = 0,
(6.2.48)
Computing the characteristic equation like in equations (6.2.20), and (6.2.21) yields
det (F Fc 1) = 0,
(6.2.49)
Because the matrix E is a diagonal and symmetric matrix, the inverse E is a diagonal and
symmetric matrix, too,
a b
a b 1 0
1 0
E 1 =
(6.2.40)
E 1 E =
=
= 1,
c d
c d 0
0 1
2
181
det
"
21
kl Fc
20
9
kl
20 2
9
kl
20 2
21
kl Fc
40
= 0,
and this nally implies the same characteristic equation like in (6.2.22),
21
21
81 2 2
kl Fc
kl Fc
k l = 0.
20
40
800
(6.2.50)
(6.2.51)
Having the same characteristic equation implies, that this problem has the same eigenvalues, i.e.
it is the same eigenvalue problem, but just another notation. With this symmetric eigenvalue
problem it is possible to compute the Rayleigh quotient,
h i pT p
pT F p
(6.2.52)
1 = R p = T +1 = T , with 1 Fc1 ,
p p
p p
Chapter 6. Exercises
182
6.3
Like in the lectures a tensor could be described by a coefcient matrix f ij , and a basis given by
ij . First do not look at the basis, just look at the coefcient matrices. In this exercise some
of the most important rules to deal with the coefcient matrices are recapitulated. The Einstein
summation convention implies for the coefcient matrix A .ji , and an arbitrary basis i.j ,
3 X
3
X
A.ji i.j ,
i=1 j=1
The dot in the superscript index of the expression A.ji shows which one of the indices is the rst
index, and which one is the second index. This dot represents an empty space, so in this case it is
easy to see, that the subscript index i is the rst index, and the superscript index j is the second
index, i.e. the index i is the row index, and the j is the column index of the coefcient matrix!
This is important to know for the multiplication of coefcient matrices. For example, what is is
the difference between the following products of coefcient matrices,
?
T
Aij B kj = Aij B jk = Di.k
, and Aij B kj =
B 11 B 12 B 13
B 21 B 22 B 23
B 31 B 32 B 33
?
C1.1 C1.2 C1.3
C2.1 C2.2 C2.3
C3.1 C3.2 C3.3
A B T = D,
T
Aij B kj Ckl = Aij B jk Ckl = Dil A B T C = D,
T lm T
A.ji Bkj Cl.k Dml = A.ji (Bjk )T C.lk
D
= Ei.m A B T C T DT = E,
T
Aij B kj uk = Aij B jk uk = vi A B T u = v,
ui B ij uj =
Aij B kj = Aim B km
Akl vl = Akj vj
, or
, etc.
? .
The Kronecker delta is dened by
Aij B kj =?
j
ji = ji = ij = ij =
B 11 B 12 B 13
B 21 B 22 B 23
B 31 B 32 B 33
D1.1 D1.2 D1.3
D2.1 D2.2 D2.3
D3.1 D3.2 D3.3
1
0
, iff
, iff
i=j
.
i 6= j
The Kronecker deltas ij , and ij are only dened in a Cartesian basis, where they represent the
metric coefcients. The other ones are dened in every basis, and in order to use the summation
convention, it is useful to prefer this notation with super- and subscript indices. Because the
Kronecker delta is the symmetric identity matrix, it is not necessary to differentiate the column
and row indices in index notation. As a rule of thumb, multiplication with a Kronecker delta
substitues an index in the same position,
vk jk = vj
j
.j
A.k
i k = A i
im s k
A i m = Ask
uT B v = .
Furthermore it is important to notice, that the dummy indices could be renamed arbitrarily,
T
[Aij ] B kj = [Aij ] B jk = Di.k
because a matrix with exchanged columns and rows is the transpose of the matrix. As a short
recap the product of a square matrix and a column matrix, resp. a (column) vector, is given by
1
u
A11 A12 A13
A11 u1 + A12 u2 + A13 u3
A u = v A21 A22 A23 u2 = A21 u1 + A22 u2 + A23 u3 Aij uj = vi .
A31 A32 A33
u3
A31 u1 + A32 u2 + A33 u3
The left-hand side of gure (6.11) sketches the rst product, and the right-hand side the second
j
Aij B jk =?
For example some products of coefcient matrices in index and matrix notation,
A1 A.2
A.3
1
1
.j
.
A.2
A.3
A = Ai = A.1
2
2
2
.2
.3
A.1
A
A
3
3
3
Aij B jk =
183
and
A.ji i.j =
v I = v,
A I = A,
A I I = A.
A B = C,
A.lk il ki = A.ii ?
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
184
This is the sum over all main diagonal elements, or just called the trace of the matrix,
tr A = A11 + A22 + A33 = tr Aji ,
and because the trace is a scalar quantity, it is independent of the basis, i.e. it is an invariant,
h i
, i.e. Aii = Aii .
tr A = tr Aji = tr Aji
For example in the 2-dimensional vector space E2 the Kronecker delta is dened by
g2 O
185
xi gij gj = xi gi
xi gij = xj .
The
gki = gik
and the
g ki = g ik
Finally this implies for the base vectors and for the coefcients or coordinates of vectors and
tensors, too. This implies, raising an index with the contravariant metric coefcients is given by
g2
gk = g ki gi
:
xk = g ki xi
, and
Aik = g ij A.k
j ,
xk = gki xi
g1
The relations between the co- and contravariant metric coefcients are given by
gk = g km gm
gk gi = g km gm gi
ik = g km gmi .
1
0
i=k
,
i 6= k
1 = g km [gmi ]
=0
=0
= 1,
= 1.
g1 g2 ,
g2 g1 ,
det g ik =
1
.
det [gik ]
, and
1
det g ik = .
g
The cross products of the Cartesian base vectors ei in the 3-dimensional Euclidean vector space
E3 are given by
gk = Akm gm .
Both sides of the equations are multiplied with g i , and nally the index i is renamed by m,
i
g ki = Akm m
If the vectors gi and gk are in the same space V, it must be possible to describe g k by a product
of gi and some coefcient like A km ,
g km = [gmi ]1
det [gik ] = g
gk gi = Akm gm gi
Than the determinants of the co- and contravariant metric coefcients are dened by
g ki = Aki
g km = Akm .
e1 e 2 = e 3 = e 3
e2 e1 = e3 = e3
, e2 e3 = e1 = e1
, e3 e2 = e1 = e1
, and
, and
e 3 e1 = e2 = e2 ,
e1 e3 = e2 = e2 .
e1 e 2 e 3
Chapter 6. Exercises
186
The cross products of the Cartesian base vectors could be described by the permutaion symbols
like this,
ei ej = eijk ek ,
1
, and eijk = ijk ,
g
and
1
ijk = eijk
g
, and eijk =
1. Simplify the index notation expressions and write down the matrix notation form.
(a) A.ji Bkj C.lk =
(b) Aij Bik Ckl =
.n .m
Dn =
(c) Cm
(b) a b + (a d) c =
4. Combine the base vectors of a general basis and simplify the expressions in index notation.
and the following relations between the Cartesian and the general permutation symbols hold
geijk
6.3.5 Exercises
(d) a (b c) =
ijk
g g = gk ,
ijk =
187
(c) (a b) (c d) =
ijk
g .
i j
j
.
eijk emnk = m
n ni m
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
(a) u v = (ui gi ) (v j gj ) =
(c) u v = (ui gi ) (v j gj ) =
(f) (u v) (w x) = [(ui gi ) (v j gj )] wk gk xl gl =
(h) (u v) (w x) = [(ui gi ) (v j gj )] wk gk xl gl =
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
188
1. Simplify the index notation expressions and write down the matrix notation form.
A.ji Bkj C.lk
A.ji
(d)
(e)
Dmn E.lm ul
AT B C = D
T
D Eu=v
uT D T E T = v T
(e) Aij B g
Am
.j
j T
(B.n
)
n
m
Am
.j
j T
(B.m
)
B kj = C lj
3. Rewrite these expressions in index notation w.r.t. a Cartesian basis, and describe what kind
of quantity the result is.
(a) (a b) c =
(b) a b + (a d) c = v
ai bj eijk ck =
ai bj eijk + (ai di ) ck = v k
(c) (a b) (c d) =
i j
j
ai bj eijk (cm dn emnk ) = ai bj cm dn eijk emnk = ai bj cm dn m
n ni m
= a m b n c m d n a n bm c m d n =
wk g k = w
ui vj wk ijk =
= ui vi wk xl klm gm = ym gm
i n
.n
(c) Aij B jk m
k = Amj B jn = Cm
n
gkn m
(f) (u v) (w x) = [(ui gi ) (v j gj )] wk gk xl gl
i j
k l
i j k l
u v gi gj w x gk gl = u v gij w x klm gm
A.k
i
kj im
A BT C = D
tr (C D) = tr E =
= (Dnm ) E.lm ul = vn
T j T
E.k = vk
ui Dj.i Ek.j = ui D.ji
jk
189
6.3.6 Solutions
(a)
ym g m = y
= ui vj wl ij nl nj il gn = ui vi wn gn ui wi vn gn = xn gn
(u v) w (u w) v = x
(h) (u v) (w x) = [(ui gi ) (v j gj )] wk gk xl gl
j im
ui v g (gm gj ) wk xl gln gk gn = um v j mjo go wk xn knp gp
= um v j mjo wk xn knp go gp
= um v j mjo wk xn knp po
= um v j wk xn mjp knp
k n
n k
= u m v j wk x n m
j m
j
(a c) (b d) (a d) (b c) =
(d) a (b c) = v
= u k v n wk x n u n v k wk x n
(u w) (v x) (u x) (v w) =
(a b) c (a c) b = v
4. Combine the base vectors of a general basis and simplify the expressions in index notation.
(a) u v = (ui gi ) (v j gj ) = ui v j gi gj = ui v j gij = ui vi =
Chapter 6. Exercises
190
6.4
scalar
p product
qk
j
g uk g
v = Tu = Tij g
= Tij uk gj gk gi
= Tij uk g jk gi
= Tij uj gi
v = Tu = vi gi
, with
vi = Tij uj .
p
R = TS = T ij gi gj S kl gk gl
x scalar y
product
= T ij S kl (gj gk ) (gi gl )
= T ij S kl gjk gi gl
= T ij Sj.l gi gl
R = TS = Ril gi gl
, with
Ril = T ij Sj.l .
p
q
= T : S = T ij gi gj : S kl gk gl
xscalar product
ij
6.4.4 Exercises
1. Compute the tensor products.
(b) TS = (Tij gi gj ) S kl gk gl =
(e) T1 = (Tij gi gj ) kl gk gl =
j i
l k
(f) 11 = i g gj k g gl =
T
(i) TTT = (Tij gi gj ) Tkl gk gl =
(b) T : S = (Tij gi gj ) : S kl gk gl =
(d) T : 1 = (Tij gi gj ) : kl gk gl =
(e) 1 : 1 = ij gi gj : kl gk gl =
kl
= T S (gi gk ) (gj gl )
= T ij S kl gik gjl
= T : S = T ij Sij .
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
191
Chapter 6. Exercises
192
6.4.5 Solutions
1. Compute the tensor products.
(f) 11 = ij gi gj kl gk gl = ij kl jk gi gl = il gi gl = ij gi gj = 1
i
j
k
l
= (Tij g g ) Tlk g g = Tij Tlk g jk gi gl = Tij Tl.j gi gl
j i
(e) 1 : 1 = i g gj : k g gl = ij kl g ik gjl = ij ji = ii = 3 = tr 1
T : TT = (Tij gi gj ) : Tkl gk gl
193
Chapter 6. Exercises
194
6.5
Deformation Mappings
(6.5.1)
(6.5.2)
The different tangent mappings of the various tangent mappings are given by
F = g i Zi
K = G i Zi
FTX = Gi gi
i
, F1
X = Gi g
i
, FT
X = g Gi ,
(6.5.4)
FT = Zi gi
i
, F1
= Zi g
, FT
= g i Zi .
(6.5.6)
KT = Zi Gi
i
, K1
= Zi G
i
, KT
= G Zi ,
(6.5.5)
identity tensor of the undeformed space identity tensor of the deformed space
1 := Zi Zi ,
(6.5.7)
1x := gi gi .
(6.5.9)
1X := Gi Gi ,
(6.5.8)
The various metric tensors of the different tangent spaces are introduced by
local metric tensor of the undeformed body
local metric tensor of the deformed body
material metric tensor of the undeformed body
- M = KT K = Gij Zi Zj , (6.5.10)
- m = FT F
- MX = 1TX 1X
- mX =
- Mx =
- mx =
(c) FX FTX =
T
(d) F1
X FX =
(e) 1X 1TX =
FTX FX
1
FT
X FX
T
1 x 1x
= gij Zi Zj , (6.5.11)
= Gij Gi Gj , (6.5.12)
i
= gij G G , (6.5.13)
= Gij gi gj , (6.5.14)
i
= gij g g
2. Compute the tensor products in index notation, and name the result with the correct name.
1
(a) KT
M K =
(b) KT MX K =
1
(c) FT
m F =
(d) FT mx F =
The identity tensors are introduced separately for the various coordinate system by
identity tensor of the parameter space
T
(b) F1
F =
(f) 1x 1Tx =
= g i Zi .
F := GRAD
,
195
6.5.2 Exercises
FX = g i Gi
. (6.5.15)
1
(e) FT
X EX F X =
(f) FTX Ex FX =
3. Compute the tensor and scalar products in index notation. Rewrite the results in tensor
notation.
(a) B = M1
m =
(b) BX =
(c) Bx =
(d) B : 1 =
(e) BT : B =
(f) BT BT : B =
(6.5.16)
(6.5.17)
(6.5.18)
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
196
6.5.3 Solutions
1. Compute the tensor products. What is represented by them?
T
1
T
1
i
j
ij
(a) K1
K = (Zi G ) (G Zj ) = G Zi Zj = M = K K
T 1
T
1
i
j
ij
(b) F1
F = (Zi g ) (g Zj ) = g Zi Zj = m = F F
T 1 1
(c) FX FTX = (gi Gi ) (Gj gj ) = Gij gi gj = M1
x = FX FX
1
T
T
1
i
j
ij
(d) F1
X FX = (Gi g ) (g Gj ) = g Gi Gj = mX = FX FX
(f) BT BT : B =
lm
ij
pr
s
k
= (G
Zl Z n
i : G gmn
prgrs Z ij Zpk) sG gjkZ Z
lm
n
= G grs G gjk p Z Zi : G gmn Zl Z
=
= (Gpr grs Gij gjp Zs Zi ) : Glm gmn Zl Zn = Gpr grs Gij gjp Glm gmn ls in
= Gpr grs Gsm gmn Gnj gjp = tr B3
2. Compute the tensor products in index notation, and name the result with the correct name.
1
(a) KT
M K =
(b) KTMX K =
(d) FTmx F =
(f) FTXEx FX =
3. Compute the tensor and scalar products in index notation. Rewrite the results in tensor
notation.
(a) B = M1
m =
(b) BX = M1
X mX =
(d) B : 1 =
197
Chapter 6. Exercises
198
6.6
199
A spiral staircase is given by the sketch in gure (6.13). The relation between the gradient angle
r
e1
e3
0 1
?e2
r
cos
= a
?
-
theorem, see also the sketch in gure (6.14), the relationship between the variable and the
variable 1 along the central line is given by
1 =
This implies
p
a 2 2 + r 2 2
i1r
, for i, r = 1, 2, 3.
Describe the forces and moments in a sectional area w.r.t. to the basis given by the moving
trihedron, with the following conditions,
i () , resp. N i = N
i ()
M = M i ai , resp. N = N i ai , with M i = M
{a1 , a2 , a3 } = {t, n, b} .
Compute the resulting forces and moments at an sectional area given by = 130 o . Consider a load vector given in the global Cartesian coordinate system by
= qr e3 ,
R
at a point S given by the angle 2 , and the radius rS . This load maybe a combination of the
self-weight of the spiral staircase and the payload of its usage.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
h
.
2
1
1 ,
=
a2 + r 2
Compute the tangent t, the normal n, and the binormal vector b w.r.t. the variable .
Determine the curvature =
, and a =
r
r
=
,
1
a2 + r 2
cos 1
= c1
r
, with
c=
1
cos
=
.
r
r 2 + a2
With this relation it is easy to see, that every expression depending on is also depending on 1 ,
this is later on important for computing some derivatives. Any arbitrary point on the central line
of the spiral staircase could be represented by a vector of position x in the Cartesian coordinate
system,
x = xi ei , and xi = xi 1 ,
or
x = x i ei
, and
xi = xi (r, ) .
Chapter 6. Exercises
200
x1 = r cos ,
x2 = r sin ,
h
1
,
x3 =
2
d
d1
b=tn
h
1
e3
2
, and = c1 .
dx
x
=
,
d1
1
= c this implies
r sin
t = a1 = r cos c
h
2
cr sin
t = a1 = cr cos .
ca
dt
da1
a1
d2 x
=
=
= 2 1,
1
1
1
d
d
d
and with this the binormal vector b is already an unit vector given by
h sin
ca sin
c
h cos b = a2 = ca cos .
b = a3 =
2
2r
cr
1
= |n |
cos
n = c2 r sin .
0
cos
n = a2 = n = sin .
0
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
, and
n =
dt
da1
d2 x
=
= 2 1,
1
1
d
d
d
, resp. a3 = a1 a2 ,
and with the denition of the cross product, represented by the expansion about the rst column,
the binormal is given by
e1 cr sin cos
ca sin
i.e. the tangent vector t is already an unit vector! For the normal unit vector n = a 2 rst the
normal vector n is computed by
n =
201
The torsion =
1
= |n | = c2 r = constant.
, with
b,1 = n =
1
n,
c2 a cos
b,1 = c2 a sin = c2 an,
0
=
1
= c2 a = constant.
Chapter 6. Exercises
202
ai,1 a =
i1r ar
| ak
a = i1r rk
cr cos
cos
a1
2
a1,1 =
= cr sin c = c r sin = c2 ra2 ,
1
0
0
the rst derivative of the second base vector a 2 is given by
sin
a2
= c cos ,
a2,1 =
1
0
and nally the rst derivative of the third base vector a 3 is given by
ca cos
cos
a3
2
= c ca sin = c a sin = c2 aa2 .
a3,1 =
1
0
0
= ai,1 a = ai,1 ak .
311
312
313
c2 a cos
cr sin
= a3,1 a1 = c2 a sin cr cos = 0,
0
ca
2
c a cos
cos
2
2
ca sin
c a cos
2
= a3,1 a3 = c a sin ca cos = 0.
0
cr
With this results the coefcient matrix of the Christoffel symbols could be represented by
0
0
c2 r
[i1r ] = c2 r 0 c2 a .
0
c2 a
0
dN =
dN + f d1 = 0,
with the rst derivative of the force vector w.r.t. to the variable 1 represented by
= a1,1 a1 = c ra2 a1 = 0,
= a1,1 a2 = c2 ra2 a2 = c2 r,
= a1,1 a3 = c2 ra2 a3 = 0,
203
sin
cr sin
211 = a2,1 a1 = c cos cr cos = c2 r,
0
ca
sin
cos
2
sin
ca sin
3
21 = a2,1 a3 = c cos ca cos = c2 a,
0
cr
N
d1 = N,1 d1 ,
1
Chapter 6. Exercises
204
b = a3 (1 )
*
n = a2 (1 )
with the computed values for the Christoffel symbols from above,
t = a1 (1 )
N,11 + 0 + N 2 c2 r + 0 + f 1 = 0,
dx = a1 d1
x = x (1 )
e3 6
R
1
f d
Q
q
e1
x-+ dx = x (1 ) + a1 d1
e2
N + dN
M + dM
Figure 6.15: An arbitrary line element with the forces, and moments in its sectional areas.
This equation rewritten in index notation, with the force vector N = N i ai given in the basis of
the moving trihedron at point P , implies
N i ai ,1 + f i ai d1 = 0,
with the chain rule,
+ f i ai = 0
+ f i ai = 0
and nally
N,11 c2 rN 2 + f 1 = 0.
N,1i ai + N i ai,1
i
N,1 ai + N i i1k ak
205
This equation system represents three component equations, one for each direction of the basis
of the moving trihedron, rst for i = 1, and k = 1, . . . , 3,
N,13 + 0 + N 2 c2 a + 0 + f 3 = 0,
N,13 c2 aN 2 + f 3 = 0.
All together the coefcient scheme of the equilibrium of forces in the basis of the moving trihedron is given by
N,11 c2 rN 2 + f 1
0
N,12 + c2 rN 1 + c2 aN 3 + f 2 = 0 .
0
N,13 c2 aN 2 + f 3
1
M + M + dM + a1 d1 (N + dN) + a1 d1 f d1 = 0,
2
1
dM + a1 d1 N + a1 dNd1 + a1 f d1 d1 = 0,
2
N,1i + N k k1i ai + f i ai = 0.
and in linear theory, i.e. neglecting the higher order terms, e.g. terms with dNd 1 , and with
d1 d1 , the equilibrium of moments is given by
N,1i + N k k1i = N i |1 ,
dM + a1 d1 N = 0.
the equilibrium condition could be rewritten in index notation only for the components,
N i |1 +f i = N,1i + N k k1i + f i = 0.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
With the rst derivative of the moment vector w.r.t. to the variable 1 given by
dM =
M 1
d = M,1 d1 ,
1
Chapter 6. Exercises
206
(M,1 + a1 N) d = 0.
The unkown forces N and moments M w.r.t. the natural basis a i should be computed, i.e.
The cross product of the rst base vector a 1 of the moving trihedron and the force vector N is
given by
a1 N = a 1 N i ai = N 1 a1 a 1 + N 2 a1 a 2 + N 3 a1 a 3
= 0 + N 2 a3 + N 3 a2 = N 2 a3 N 3 a2 ,
a1 N = N 2 a3 N 3 a 2 ,
because the base vectors ai form an orthonormal basis. The following steps are the same as the
ones for the force equilibrium equations,
M,1 + N 2 a3 N 3 a2 d1 = 0,
N = N i ai
and is acting at a point S, given by an angle 2 , and a radius rS , see also the free-body diagram given by gure (6.16). The free-body diagram and the load vector are given in the global
r
e1
M,11 + 0 + M 2 c2 r + 0 = 0,
M,11 c2 rM 2 = 0,
?e2
1
R
M,13
+M
113
213
2
313
M,13 c2 aM 2 + N 2 = 0.
All together the coefcient scheme of the equilibrium of moments in the basis of the moving
trihedron is given by
M,11 c2 rM 2
0
2
2
1
2
3
3
M,1 + c rM + c aM N = 0 .
0
M,13 c2 aM 2 + N 2
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
2
N
1
3 N
?N
1
M
3 2
R
R
?
+M
+M
+ N = 0,
M,13 + 0 + M 2 c a + 0 + N 2 = 0,
?
?
3
M
rS
M,12 + c2 rM 1 + c2 aM 3 N 3 = 0.
and nally in the direction of the third base vector a 3 , i.e. i = 3, and k = 1, . . . , 3,
2
M
e3
M,1i + M k k1 ai + N 2 a3 N 3 a2 = M i |1 ai + N 2 a3 N 3 a2 = 0.
Again this equation system represents three equations, one for each direction of the basis of the
moving trihedron, rst for i = 1, i.e. in the direction of the base vector a 1 , and k = 1, . . . , 3,
M = M i ai .
f = fi ei ,
and nally
, and
M,1i ai + M i ai,1 + N 2 a3 N 3 a2 = 0,
207
, with
1
R
0
2 = 0 .
R
3
R
qr
First the equilibrium conditions of forces in the directions of the base vectors e i of the global
Cartesian coordinate system are established,
X
1 + R
1 = 0 N
1 = 0,
F e1 = 0 N
3 + R
3 = 0 N
3 = R
3.
F e3 = 0 N
2 + R
2 = 0 N
2 = 0,
F e2 = 0 N
Chapter 6. Exercises
208
Than the equilibrium conditions of moments in the directions of the base vectors e i of the global
Cartesian coordinate system w.r.t. the point T are established
X
3 r sin + rS sin + M
1 = 0
Me(T1 ) = 0 R
2
1 = R
3 r sin
M
rS sin
,
2
2
(T )
3
rS cos + r cos
2 = 0
R
Me 2 = 0
+M
2
2
2 = R
3 rS cos r cos ,
M
2
2
X
3 = 0.
Me(T3 ) = 0 M
Finally the resulting equations of equilibrium are given by
3
r sin rS sin
0
R
2
2
3 rS cos r cos
= 0 , and M
= R
.
N
2
2
3
R
0
Now the problem is, that the equilibrium conditions are given in the global Cartesian coordinate
system, but the results should be descirbed in the basis of the moving trihedron. For this reason it
i ei , into N = N i ai ,
i ei , and M = M
is necessary to transform the results from above, i.e. N = N
and M = M i ai ! The Cartesian basis ei should be transformed by a tensor S into the basis ai ,
S = S rs er es
ai = Sei = S rs si er = S.ir er ,
[S.ir ]T
[er ]
, with
see also the denitions for the base vectors of the moving trihedron in the sections above! Then
the retransformation from the basis of the moving trihedron into the global Cartesian basis should
be given by S1 = T, with
ei = Tai , with T = T.sr ar as ,
ei = (T.sr ar as ) ai = T.sr is ar = T.ir ar .
Comparing this transformation relations implies
ai = S.ir er = S.ir T.rm am
k
,
ik = S.ir T.rm m
| ak ,
ik = T.rk S.ir ,
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
209
T.rk = [S.ir ]1 .
Because both basis, i.e. ei , and ai , are orthonormal basis, the tensor T must describe an orthonormal rotation! The tensor of an orthonormal rotation is characterized by
T1 = TT
, resp.
T = S1 = ST ,
[S.ir ]1
[S.ir ]T
cr sin cr cos ca
0 .
= cos sin
ca sin ca cos cr
With this the relations between the base vectors of the different basis could be given by
ai = S.ir er
, resp. er = T.rk ak ,
T
T.rk
cr sin cos ca sin
a1
[ak ] = cr cos sin ca cos a2 ,
a3
ca
0
cr
i ei , into
i ei , and M = M
With this it is easy to compute the nal results, i.e. to transform N = N
N = N i ai , and M = M i ai . With the known transformation ei = T.ik ak the force vector could
be represented by
i T k ak = N k ak .
i ei = N
N=N
.i
Comparing only the coefcients implies
i,
N k = T.ik N
and with this the coefcients of the force vector N w.r.t. the basis of the moving trihedron in the
sectional area at point T are given by
1 3
1
ca
R
cr sin cr cos ca
0
N
N
N 2 = cos sin
0 0 N 2 = 0 .
3 cr
3
R
ca sin ca cos cr
R
N3
N3
By the analogous comparison the coefcients of the moment vector M w.r.t. the basis of the
moving trihedron in the sectional area at point T are given by
1
1
1
1 sin + M
2 cos
cr M
cr sin cr cos ca
M
M
M
2
2
2
1
2
M = M
cos + M
sin .
M = cos sin
0 M
2 cos
1 sin + M
ca sin ca cos cr
M3
0
M3
ca M
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
210
6.7
211
ij
The given cylinderical shell is described by the parameter lines i , with i = 1, 2, 3. The relations
= gi g j
, with
ij
6 0 2
= 0 0 0 .
2 0 5
g2
n=
n
g1
Compute the covariant base vectors gi and the contravariant base vectors g i at the point P .
e1
(0, 0, 0)
(8, 0, 4)
x
-
e3
e2
(0, 2, 0)
3
(8, 2, 0)
3
(8, 0, 0)
2
g1 + g 3
= n r er .
|g1 + g3 |
[cm]
between the Cartesian coordinates and the parameters, i.e. the curvilinear coordinates, are given
by the vector of position x,
x1 = 5 2 sin
1 ,
a
x2 = 3 , and
x3 = 5 2 cos
1 ,
a
Work out the physical components from the contravariant stress tensor.
Determine the invariants,
Determine the coefcients of the tensor w.r.t. the basis in mixed formulation, gi gk ,
and w.r.t. the Cartesian basis, (ei ek ).
I = tr ,
1
II =
(tr )2 tr ()2 ,
2
III = det ,
for the three different representations of the stress tensor .
Calculate the principal stresses and the principal stress directions.
Compute the specic deformation energy W spec = 12 : at the point P , with
=
1
(gik ik ) gi gk .
100
8
2
1 = ; 2 = ; 3 = 2 ,
10
10
Determine the stress vector tn at the point P w.r.t. the sectional area given by the normal
vector n. Furthermore calculate the normal stress t , and the resulting shear stress tk for
this direction.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
212
gi = g ik gk ,
and for i = 1, . . . , 3 the relations between co- and contravariant base vectors are given by
x
x
=
ei .
k
k
With this denition the covariant base vectors are computed by
x
cos
1 e1 + 0 e 2
sin
1 e3 ,
g1 =
= 5 2
+ 5 2
1
a
a
a
a
x
= sin
+ 0 e2
+ cos
1 e1
1 e3 ,
g2 =
2
a
a
x
g3 =
= 0 e1
+ (1) e2 + 0 e3 ,
3
and nally the covariant base vectors of the curvilinear coordinate system are given by
cos a
sin a 1
0
1 ,
0
0
,
g
,
and
g
=
=
g1 = 5
2
3
a
0
sin a 1
cos a 1
g 1 = b 2 g1
gk =
or by
cos c
1
0
g1 =
b
sin c
with the abbreviations
b=
, g2 =
5
a
=
(5 2 )
3
sin c
0
cos c
213
, and g3 = 1 ,
0
, and c =
= .
a
10
In order to determine the contravariant base vectors of the curvilinear coordinate system, it is
necessary to multiply the covariant base vectors with the contravariant metric coefcients. The
contravariant metric coefcients g ik could be computed by the inverse of the covariant metric
coefcients g ik ,
1
[gik ] = g ik
.
So the rst step is to compute the covariant metric coefcients g ik ,
1
0 0
b2
gik = gi gk , i.e. [gik ] = 0 1 0 .
0 0 1
The relationship between the co- and contravariant metric coefcients is used in its inverse form,
in order to compute the contravariant metric coefcients,
2
b 0 0
ik
ik
1
g = [gik ]
, resp.
g = 0 1 0 .
0 0 1
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
cos c
g1 = b 0
sin c
or in detail
, g 2 = g2
, and g3 = g3 ,
sin c
, g2 = 0
cos c
1
cos a
(5 2 )
g =
0
a
sin a 1
1
g =
0
g3 = 1 ,
0
, and
sin
cos
0
a
0
, and g = 1 .
0
3
At the given point P the co- and contravariant base vectors g i , and gi , are given by
cos 10
cos 10
3
5
0 ,
0 , g1 =
g1 =
5
3
sin 10
sin 10
sin 10
2
0 , and
g2 = g =
cos 10
0
g3 = g3 = 1 .
0
, and
gm = gmk gk ,
= im gmk gi gk = ik gi gk ,
ik
im
1
0 0
6 0 2
b2
[gmk ] = 0 0 0 0 1 0 .
0 0 1
2 0 5
Chapter 6. Exercises
214
215
Solving this matrix product implies the coefcient matrix [ ik ] w.r.t. the basis gi gk , i.e. the
stress tensor w.r.t. to the mixed basis is given by
6
0 2
b2
i
i
k
= k gi g , with
k = 0 0 0 ,
2
0 5
b2
Solving this matrix product implies the coefcient matrix of the stress tensor w.r.t. the Cartesian
basis, i.e. the stress tensor w.r.t. the Cartesian basis is given by
6
cos2 c
2b cos c b62 sin c cos c
b2
5
2b sin c ,
=
ik ei ek , with [
rs ] = 2b cos c
6
2
6
sin
c
cos
c
sin
c
sin2 c
2
b
b
b2
The relationships between the Cartesian coordinate system and the curvilinear coordinate system
are described by
and nally at the point P the stress tensor w.r.t. the mixed basis is given by
542
0 2
25
i
i
k
0 0 0 .
= k gi g , with
k =
18 2
0 5
25
1
cos c sin c 0
b
T k
0
1 ,
, with [Bki ] = 0
[gi ] = [Bki ] e
1
sin c cos c
0
b
see also the denition of the covariant base vectors above. The stress tensor w.r.t. the Cartesian
basis is computed by
ik
= gi g k
, and gi = Bri e
, resp. gk = Bsk e ,
= ik Bri Bsk er es ,
g(k)(k)
ik = ik p
,
g (i)(i)
see also the lecture notes. The physical components ik of a tensor = ik gi gk consider,
that the base vectors of an arbitrary curvilinear coordinate system do not have to be unit vectors!
In Cartesian coordinate systems the base vectors ei do not inuence the physical value of the
components of the coefcient matrix of a tensor, because the base vectors are unit vectors and
orthogonal to each other. But in general coordinates the base vectors do inuence the physical
value of the components of the coefcient matrix, because they are in general no unit vectors,
and not orthogonal to each other. Here the contravariant stress tensor is given by
6 0 2
ik = 0 0 0 .
= ij gi gk , with
2 0 5
In order to compute the physical components of the stress tensor , it is necessary to solve the
denition given above. The numerator and denominator of this denition are given by the square
roots of the co- and contravariant metric coefcients g (i)(i) , and g (i)(i) , i.e.
ik
rs = Bri Bsk ,
the coefcient matrix of the stress tensor w.r.t. the Cartesian basis is dened by
1
1
cos c sin c 0
cos c 0
6 0 2
b
b
ik
T
0
1 0 0 0 sin c 0
[
rs ] = [Bri ] [Bsk ] = 0
1
0
1
2 0 5
sin c cos c
0
b
and nally at the point P the stress tensor w.r.t. the Cartesian basis is given by
542
54 2
cos2 10
6
cos 10
sin 10
cos 10
25
5
25
6
5 cos 10
5
5 sin 10 .
rs ] =
=
ik ei ek , with [
54 2
54 2
sin 10 cos 10 5 sin 10
sin2 10
25
25
sin c
cos c .
0
1
b
1
b
=b
g11 =
g 11
g22 = 1
p
g 22 = 1
g33 = 1,
p
g 33 = 1.
Finally the coefcient matrix of the physical components of the contravariant stress tensor =
ik gi gk is given by
6
0 2b
h i
b2
g(k)(k)
ik
ik
ik
= p
, and
= 0 0 0 .
2
g (i)(i)
0 5
b
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
216
6.7.5 Invariants
217
tr 2 = T : = ki gi gk : ( rs gr gs ) = ki rs gir gks = kr rk .
The stress tensor could be described w.r.t. the three following basis, i.e.
= ik gi gk , w.r.t. the covariant basis of the curvilinear coordinate system,
= ik gi gk , w.r.t. the mixed basis of the curvilinear coordinate system, and
=
ik ei ek , w.r.t. the basis of the Cartesian coordinate system.
The rst invariant I of the stress tensor is dened by the trace of the stress tensor, i.e.
I = tr = : 1.
The rst invariant w.r.t. the covariant basis of the curvilinear coordinate system is given by
I = tr = ik gi gk : gml gl gm = ik gml il g km = ik il kl = ii ,
6
I = tr = ii = 2 + 5.
b
Finally the second invariant w.r.t. the covariant basis of the curvilinear coordinate system
is given by
1
(tr )2 tr 2
2"
2
#
36
8
1 36 60
36
8
1
6
+
5
+
+
25
=
+
+
25
25
,
=
2
b2
b4
b2
2 b4
b2
b4
b2
II =
II =
2 = ik gi gk ( rs gr gs ) = ik rs rk gi gs = ik ks gi gs ,
and then
tr 2 = 2 : 1 = ik ks gi gs : gml gl gm = ik ks gml il g sm = ik ki ,
The rst invariant w.r.t. the mixed basis of the Cartesian coordinate system is given by
I = tr = (
ik ei ek ) : (ml em el ) =
ik ml im kl =
ii ,
6
ii = 2 + 5.
I = tr =
b
The second invariant II of the stress tensor is dened by the half difference of the trace to the
second of the stress tensor, and the trace of the stress tensor to the second, i.e.
2 = ik gi gk ( rs gr gs ) = ik rs gkr gi gs ,
and then
tr = : 1 = gkr gi gs : gml gl gm
2
ik rs
ik rs
gkr gml il sm =
2
6
tr 2 =
b2
2
+2 2 2
b
ik rs
gkr gsi =
+ 25 =
tr 2 = T : = ki gi gk : ( rs gr gs ) = ki rs gir g ks = sr rs .
This intermediate result is the same like above, and the rst invariants, i.e. the trace tr
and (tr )2 , too, are equal for the different basis, i.e. all further steps will be the same like
above. Combining all this nally implies, that the second invariant w.r.t. the mixed basis
of the curvilinear coordinate system is given by
II =
1
II =
(tr )2 tr 2 .
2
26
.
b2
ir ri
36
8
+ 2 + 25,
b4
b
26
b2
, too.
Chapter 6. Exercises
218
36
4
36 2
4
36
2
2
sin
c
cos
c
+
sin
c
+ 4 sin4 c + 25
tr 2 = 4 cos4 c + 2 2 cos2 c + 2
b
b
b4
b2
b
8 2
36 4
2
2
4
2
= 4 cos c + 2 sin c cos c + sin c + 2 cos c + sin c + 25
b
b
36
8
tr 2 = 4 + 2 + 25,
b
b
i.e. the further steps are the same like above. And with this nally the second invariant
w.r.t. the basis of the Cartesian coordinate system is given by
II =
26
b2
, too.
The third invariant III of the stress tensor is dened by the determinant of the stress tensor, i.e.
III = det =
1 T T
1
1
(tr )3 (tr ) tr 2 +
: .
6
2
3
In order to compute the third invariant it is necessary to solve the three terms in the summation of the denition of the third invariant. The rst term is given by
1
1 6
1 216 540 450
(tr )3 =
+
5
=
+
+
+
125
6
6 b2
6 b6
b4
b2
36
90
75
125
1
(tr )3 = 6 + 4 + 2 +
,
6
b
b
b
6
and the second term is given by
1
36
8
1 6
(tr ) tr 2 =
+
5
+
+
25
2
2 b2
b4
b2
1
108
114
95
125
(tr ) tr 2 =
.
+
+
+
2
b6
b4
b2
2
The third term is not so easy to compute, because it does not include the trace, but a scalar
product and a tensor product with the transpose of the stress tensor, i.e.
1
1 T T
: = T :
3
3
, with
T = T T = ()T .
or in index notation
T T
: = ki gi gk ( sr gr gs ) : lm gl gm
lm
ki s
ki sr
= gkr gi gs : gl gm = k gi gs : lm gl gm
= ki sk lm gil gsm = kl sk ls ,
: = kl ls sk .
219
T = T T = ki gi gk ( sr gr gs ) = ki sr gkr gi gs
si gi gs = ki sk gi gs = sk ki gi gs ,
36
6
0 2
+4 0
6 0 2
b2
b2
0
[si ] = 0 0 0 0 0 0 = 0
2
12
0 5
+ 10 0
2 0 5
b2
b2
12
b2
4
b2
+ 10
0 .
+ 25
T : = T T : = si gi gs : lm gl gm ,
T : = si lm gil gsm = si gil ls = sl ls ,
rst the coefcient matrix of the tensor w.r.t. the mixed basis is computed by
[ sl ] = si [gil ] ,
1
36
36
+ 4 0 12
+ 10
0 0
+ b42 0 12
+ 10
b2
b2
b2
b4
b2
s
0
0 0 1 0 = 0
0
0 ,
[ l ] = 0
12
12
+ 10 0 b42 + 25
+ 10
0 b42 + 25
0 0 1
b2
b4
b2
and then the nal result for the third term is given by
1
1
1 T T
1
: = T : = kl ls sk = sl ls
3
3
3
3
36
4
1
4 6
12
2
12 10
+
5
=
+
+
+
10
+
2
+
+
25
3
b4
b2 b2
b2
b2
b4
b2
b2
1 T T
72 24 20 125
.
:= 6 + 4 + 2 +
3
b
b
b
3
Then the complete third invariant w.r.t. the covariant basis of the curvilinear coordinate
system is given by
1 T T
1
1
:
III = det = (tr )3 (tr ) tr 2 +
6
2
36 90 75 125
108 114 95 125
=
+
+
+
+
+
+
b6
b4
b2
6
b6
b4
b2
2
72 24 20 125
+ 4 + 2 +
+
b6
b
b
3
1
1
1
= 6 (36 108 + 72) + 4 (90 114 + 24) + 2 (75 95 + 20)
b
b
b
+
6
2
3
III = det = 0.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
220
For the third invariant w.r.t. the mixed basis of the curvilinear coordinate system the rst
and second term are already known, because all scalar quantities, like tr , are still the
same, see also the rst case of determining the third invariant. It is only necessary to have
a look at the third term given by
T T
: = ik gk gi ( rs gs gr ) : lm gl gm
i r s k
l
= k s i g gr : m gl gm = sk rs gk gr : lm gl gm
= sk rs lm lk rm = sl ms lm ,
T T : = sl lm ms ,
i.e. this scalar product is the same like above. With all three terms of the summation given
by the same scalar quantities like above, the third invariant w.r.t. the mixed basis of the
curvilinear coordinate system is given by
The third invariant w.r.t. the basis of the Cartesian coordinate system is given by
III = det = 0.
The nal result is, that the invariants of every arbitrary tensor could be computed w.r.t. any
basis, and still keep the same, i.e.
6
I = tr
=:1
= 2 + 5,
b
1
1
26
II =
(tr )2 tr 2 =
( : 1)2 T :
= 2,
2
2
b
1 T T
1
1
= (tr )3 (tr ) tr 2 +
III = det
:
6
2
3
1 T T
1
1
= ( : 1)3 ( : 1) T : +
: = 0.
6
2
3
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
, and
n 0 = n r er = n r er ,
III = det = 0.
The third case, i.e. the third invariant w.r.t. to the basis of the Cartesian coordinate system,
is very easy to solve, because in a Cartesian coordinate system it is sufcient to compute
the determinant of the coefcient matrix of the tensor! For example the determinant of the
coefcient matrix
rs is expanded about the rst row
6
2 cos2 c
2b cos c b62 sin c cos c
b 2
5
2b sin c
det = det [
rs ] = b cos c
6
62 sin c cos c 2 sin c
sin2 c
b
b2
b
30 2
4
2
12
6
12
sin c 2 sin2 c + cos c 3 sin2 c cos c + 3 sin2 c cos c
= 2 cos2 c
2
b
b
b
b
b
b
6
4
30
+ 2 sin c cos c 2 sin c cos c 2 sin c cos c
b
b
b
156 2
156 2
2
= 4 sin c cos c 0 4 sin c cos2 c
b
b
det = det [
rs ] = 0.
221
ik nk ei = nk ek | el ,
ik nk il = nk kl ,
lk nk = nk lk ,
and nally the characteristic equation is given by
(
lk lk ) nk = 0,
det (
ik ik ) = 0.
The characteristic equation of the eigenvalue problem could be rewritten with the already known
invariants, i.e.
det (
ik ik ) = III II + I 2 3 = 0,
I + II III = 0,
6
26
3
2 + 5 2 + 2 0 = 0,
b
b
6
26
2 2 + 5 + 2 = 0,
b
b
3
this implies the rst eigenvalue, i.e. the rst principal stress,
1 = 0.
The other eigenvalues are computed by solving the quadratic equation
6
26
2 2 + 5 + 2 = 2 d + e = 0,
b
b
r
s
2
1 6
26
1
1 6
d2
+
5
+
5
=
2/3 =
d
e,
2
2
2
2 b
4 b
b
2
4
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Chapter 6. Exercises
222
this implies the second and third eigenvalue, i.e. the second and third principal stress,
r
r
d2
d2
6
1
1
26
e , and 3 = d
e , with d = 2 + 5 , and e = 2 .
2 = d +
2
4
2
4
b
b
sin c
n1 = 0 .
cos c
In order to compute the principal directions the stress tensor w.r.t. the curvilinear basis, and a
normal unit vector, i.e.
= ik gi gk , and n = nr gr ,
are used, then the eigenvalue problem is given by
n = n,
and the left-hand side is rewritten in index notation,
The eigenvalue problem in index notation w.r.t. the curvilinear basis is given by
ik nk gi = ni gi = nk ki gi ,
i
i
k k nk gi = 0,
i
k ik nk = 0,
b2
i 0
2
ni1
0 ni2 = 0.
0
i
2
0 5 i
ni3
b2
Combining the rst and the last row of this system of equations yields an equation to determine
the coefcient n i3 depending on the associated principal stress i , i.e.
4
6
6
26
(5 i ) 2 i 2 ni3 = 2 i 5 + 2 + 2i ni3 = e i d + 2i ni3 = 0.
b
b
b
b
The coefcient n i2 could be computed by the second line of this system of equations, i.e.
i ni2 = 0,
and then the coefcient n i1 depending on the associated principal stress i and the already known
coefcient n i3 is given by
2ni3
ni1 = 6
.
i
2
b
The associated principal direction to the principal stresses are computed by inserting the values
i in the equations above.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
d2
4
e implies
1
n11 = b2 (5 2 ) = ,
2
n2 = n1 g1 + n2 g2 + n3 g3 = g1 + 0 + g3 ,
b cos c
1
1 , with = b2 (5 2 ) .
n2 =
2
b sin c
n13 = R
n = ik gi gk nr gr = ik nr gkr gi = ir nr gi .
223
n12 = 0
d2
4
e implies
b cos c
n3 = 1
b sin c
, with
1
= b2 (5 3 ) .
2
, with
= ik gi gk
, and
1
(gik ik ) gi gk ,
100
1
1 lm
1
(gik ik ) gi gk
gl g m :
Wspec = : =
2
2
100
1 ik
1 lm
i k
(gik ik ) l m =
(gik ik )
=
200
200
1 ik
=
gik ik ,
200
1 i
1
Wspec = : =
i ii ,
2
200
because the Kronecker delta ik is given w.r.t. to the Cartesian basis, i.e.
ik = i k = ik = ik .
Chapter 6. Exercises
224
With the trace of the stress tensor w.r.t. to the mixed basis of the curvilinear coordinate system,
and the trace of the stress tensor w.r.t. to the covariant basis of the curvilinear coordinate system
given by
6
ii = 2 + 5 , and ii = 11,
b
the specic deformation energy is given by
1
6
1
Wspec = : =
6
,
2
200 b2
(6 + 2b2 ) cos c
+ 2 cos c
b2
1
1
2
5b =
2b 5b3 .
tn =
b2 + 1 6 +b 2 sin c
b2 b2 + 1 (6 + 2b2 ) sin c
1
200
54 2
6
25
n=
and this implies with
cos c
g1 + g3 = 1
1
sin c
b
1
nally
, and
t = n
= tn n =
|g1 + g3 | =
, with
= |t | = tn n,
(6 + 2b2 ) cos c
cos c
1
1
3
b
2b 5b
= tn n =
b2 b2 + 1 (6 + 2b2 ) sin c
b2 + 1 sin c
g1 + g 3
,
|g1 + g3 |
tk = t n t .
0.0766.
b2
1
+ 1,
b2
cos c
1
b = nr er = nr er .
n=
b2 + 1 sin c
=
ik ei ek ,
the stress vector tn at the point P is given by
ik ei ek nr er =
ik nr kr ei =
ik nk ei ,
tn = n =
5b4 + 4b4 + 6
.
b2 (b2 + 1)
cos c
5b4 + 4b2 + 6
b ,
t = n =
b2 (b2 + 1) b2 + 1 sin c
(6 + 2b2 ) cos c
cos c
4
2
+
4b
+
6
5b
1
3
2b 5b
b ,
tk = t n t =
b2 b2 + 1 (6 + 2b2 ) sin c
b2 (b2 + 1) b2 + 1 sin c
cos c
4 3b2
1
.
tk =
(b2 + 1) b2 + 1 sinb c
ti =
ik nk .
6
b2
cos c
sin c cos c
1
2
b ,
b sin c
b2 + 1 sin c
6
sin2 c
b2
225
226
Chapter 6. Exercises
Appendix A
Formulary
A.1
A.1.1 Basis
ei orthonormal cartesian base vectors En
gi covariant base vectors
En
En
metric coefcients
gik = gi gk
gi = gik gk
g ik = gi gk
ki
ki
(ik
gi = g ik gk
gi = ki gk
ei = ki ek
(ei = ik ek )
= g gk
= e ek
= ei ek )
= T ik gi gk = Ti k gi gk
227
Appendix A. Formulary
228
= ui gi v k gk wm gm = v k gk wm gm ui gi
k m
i
= v w gk m u g i = v k w k u i g i
T = T ik gi gk
T w = T ik gi gk wm gm = T ik wk gi
mit
u=1u
1 = gi gi = gj gj = ij gj gi
= g ij gi gj = gij gi gj
u = gi gi uk gk = uk g ik gi = uk gk = ui gi = u
u=Aw
u (T v) = v TT u
T = T ik gi gk
with
TT = T ik gk gi = T ki gi gk
and
(u v)T = v u
(A B)T = BT AT
C = A B = Aik gi gk (Bmn gm gn )
und
with
T ik = gi T gk
T
ik
im kn
= g g Tmn
Tik = gi (T gk )
Tki
im
= g Tmk
AT
=A
Tki = gi (T gk )
;
;
etc.
det Q = 1
vv =uu
tr T = T : 1 = T ik gi gk : (gm gm ) = T ik gim km
tr (a b) := a b
229
resp.
= T ik gik = Tii
tr (A B) = A : BT
or
A B = tr A BT = tr BT A
tr (A B) = tr (B A) = B : AT
etc.
transformation gi gi ; gk gk
gi = 1 gi = gk gk gi = gk gi gk = Aki gk
Appendix A. Formulary
230
g i = A gi
with
A = gk gm gk gm = Akm gk gm
k
gi = g gi gk = Aki gk
g i = A gi
with
gi = (gk gi ) gk = Aki gk
i
A = (gk gm ) g g = Akm g g
Ami Bmk = ik
Bki gk
gi = B gi
with
B = (gk gm ) gk gm = Bkm gk gm
k
i
i
i k
g = gk g g = B k g
gi = A g i
with
k
k
gi = g g i g k = A i g k
gi = A g i
A = (gm gk ) g g = Amk g g
gi = B gi
with
B = (gk gm ) gk gm = B k gk gm
i
gi = gk gi gk = B k gk
km
gi = B gi
with
B = gk gm gk gm = B gk gm
ki
gi = gk gi gk = B gk
Bki vk
v =
ki
= B vk
vi = A i vk = Aki vk
ki
vi = B k vk = B vk
i.e. the coefcients of the vector components transform while changing the coordinate systems
like the base vectors themselves.
i
gi = 1 gi = gk gk gi = gk gi gk = B k gk
Ami A m = ik
k
A = g k gm g k g m = A m g k g m
k
Bmk = A m
etc.
k
gi = 1 g i = g k g k g i = g k g i g k = A i g k
Bmk = A m
v = v i g i = v i gi = v i g i = v i g i
inverse relations gi gi ; gk gk
AA=1
B m Bkm = ki
etc.
gi = B gi
with
B = gk gm gk gm = B km gk gm
gi = gk gi gk = B ki gk
with
BB=1
or
Bmi B k = ki
Furthermore
g = 1 g = g g k g = gk g g =
gi = (gk gi ) gk = Aki gk
231
or
etc.
AA=1
m
A i Akm = ik
T = T ik gi gk = T ik gi gk = Tik gi gk = Ti k gi gk
ik
= T gi gk = T k gi gk = T ik gi gk = T i gi gk
ik
i
k
= A m A n T mn
=
T ik =
A m Ank T m
n
Ami Ank Tmn
,
,
,
i.e. the coefcients of the tensor components transform like the tensor basis. The tensor basis
transform like the base vectors.
TU Braunschweig, CSE Vector and Tensor Calculus 22nd October 2003
Appendix A. Formulary
232
T ik
(T 1) = 0
ki xi = 0
!
characteristic polynomial
det (T 1) = 0
det T ik ki xi = 0
A.2
233
f () = I3 I2 + 2 I1 3 = 0
If T = TT ; Tki R, then the eigenvectors are orthogonal and the eigenvalues are real. invariants
of a tensor
I1 = 1 + 2 + 3 = tr T = T ik
1
I2 = 1 2 + 2 3 + 3 1 =
(tr T)2 tr T2
2
1 i k
=
T i T k T ik T ki
2
I3 = 1 2 3 = det T = det T ik
v = v (x)
T = T (x)
A vector eld v = v (x) is differentiable in x, if a linear mapping L (x) exists, such that
, if |y| 0.
v (x + y) = v (x) + L (x) y + O y2
The mapping L (x) is called the gradient or Frechet derivative v 0 (x), also represented by the
operator
L (x) = grad v (x) .
rules
rules
(k) = ikm gi gm
Appendix A. Formulary
234
( )k
= ( )k,i
with
(k) = grad (gk ) = gk,i gi
i
s
gi,k = (i) gk = il gs gl gk = sik gs
;
gi,k gs = sik
ikl =
i
g,k
= isk gs
1
etc.
= (gkl,i + gil,k gik,l )
2
ZA
etc.
gls sik
ei,k = 0
235
u nda =
ui ni da =
T nda =
Ti k nk gi da =
with n
A
V
div udV
ZV
ZV
ZV
V
ui |i dV
div TdV
Ti k |k gi dV
normal vector of the surface element
surface
volume
v
(v i gi )
gk =
gk
k
k
v i
gi
= k gi g k + v i k g k
| {z }
grad v (x) =
(i)
i
= v,k
gi gk + v i (i)
i
= v,k
gi gk + v i gi,k gk
i
grad v (x) = v,k + v s isk gi gk = v i |k gi gk
T k (T ij gi gj ) k
g =
g
k
k
gi
gj
= T,kij (gi gj ) gk + T ij
gj gk + T ij gi k gk
k
div T (x) =
236
Appendix A. Formulary
Appendix B
Nomenclature
Notation
, , , . . .
a, b, c, . . .
aT , bT , cT , . . .
A, B, C, . . .
a, b, c, . . .
A,
3 C,
3 ...
3 B,
A, B, C, . . .
A, B, C, . . .
Notation
tr
det
sym
skew
dev
grad =
div
rot
Description
scalar quantities in R
column matrices or vectors in Rn
row matrices or vectors in Rn
matrices in Rn Rn
vectors or rst order tensors in E n
second order tensors in En En
third order tensors in En En En
fourth order tensors in En En En En
Description
the trace operator of a tensor or a matrix
the determinant operator of a tensor or a matrix
the symmetric part of a tensor or a matrix
the antisymmetric or skew part of a tensor or a matrix
the deviator part of a tensor or a matrix
the gradient operator
the divergence operator
the rotation operator
the laplacian or the Laplace operator
237
Appendix B. Nomenclature
238
Bibliography
[1] Ralph Abraham, Jerrold E. Marsden, and Tudor Ratiu. Manifolds, Tensor Analysis and
Applications. Applied Mathematical Sciences. Springer-Verlag, Berlin, Heidelberg, New
York, second edition, 1988.
[2] Albrecht Beutelspacher. Lineare Algebra. Vieweg Verlag, Braunschweig, Wiesbaden,
1998.
Notation
R
R3
E3
E3 E 3
{e1 , e2 , e3 }
{g1 , g2 , g3 }
{g1 , g2 , g3 }
gij
g ij
g = g ij gi gj
Description
the set of the real numbers
the set of real-valued triples
the 3-dimensional Euclidean vector space
the space of second order tensors over the Euclidean vector space
3-dimensional Cartesian basis
3-dimensional arbitrary covariant basis
3-dimensional arbitrary contravariant basis
covariant metric coefcients
contravariant metric coefcients
metric tensor
[3] Reint de Boer. Vektor- und Tensorrechnung fr Ingenieure. Springer-Verlag, Berlin, Heidelberg, New York, 1982.
[4] Gerd Fischer. Lineare Algebra. Vieweg Verlag, Braunschweig, Wiesbaden, 1997.
[5] Jimmie Gilbert and Linda Gilbert. Linear Algebra and Matrix Theory. Academic Press,
San Diego, 1995.
[6] Paul R. Halmos. Finite-Dimensional Vector Spaces. Undergraduate Texts in Mathematics.
Springer-Verlag, Berlin, Heidelberg, New York, 1974.
[7] Hans Karl Iben. Tensorrechnung. Mathematik fr Ingenieure und Naturwissenschaftler.
Teubner-Verlag, Stuttgart, Leipzig, 1999.
[8] Klaus Jnich. Lineare Algebra. Springer-Verlag, Berlin, Heidelberg, New York, 1998.
[9] Wilhelm Klingenberg. Lineare Algebra und Geometrie. Springer-Verlag, Berlin, Heidelberg, New York, second edition, 1992.
[10] Allan D. Kraus. Matrices for Engineers. Springer-Verlag, Berlin, Heidelberg, New York,
1987.
[11] Paul C. Matthews. Vector Calculus. Undergraduate Mathematics Series. Springer-Verlag,
Berlin, Heidelberg, New York, 1998.
[12] James G. Simmonds. A Brief on Tensor Analysis. Undergraduate Texts in Mathematics.
Springer-Verlag, Berlin, Heidelberg, New York, second edition, 1994.
[13] Erwin Stein. Unterlagen zur Vorlesung Mathematik V fr konstr. Ingenieure Matrizenund Tensorrechnung SS 94. Institut fr Baumechanik und Numerische Mechanik, Universitt Hannover, 1994.
239
240
Bibliography
[14] Rudolf Zurmhl. Matrizen und ihre technischen Anwendungen. Springer-Verlag, Berlin,
Heidelberg, New York, fourth edition, 1964.
- Integralnorm, L1-Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
L2-norm
- L2-Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
l1-norm
- Summennorm, l1-Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
n-tuple
- n-Tupel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
p-norm
- Maximumsnorm, p-Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
absolute norm
- Gesamtnorm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
absolute value
- Betrag . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
additive
- additiv . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
additive identity
additive inverse
afne vector
- afner Vektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
antisymmetric
- schiefsymmetrisch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
antisymmetric matrix
- scheifsymmetrische Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . 41
antisymmetric part
- antisymmetrischer Anteil . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
- Flchenvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
associative
- assoziativ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
associative rule
- Assoziativgesetz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
associative under matrix addition assoziativ bzgl. Matrizenaddition . . . . . . . . . . . . . . . . . . . . . 42
base vectors
- Basisvektoren . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31, 87
basis
- Basis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
241
242
243
bijective
- bijektiv . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
complex numbers
bilinear
- bilinear . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
bilinear form
- Bilinearform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
- komplexe Zahlen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
binormal unit
- Binormaleneinheitsvektor . . . . . . . . . . . . . . . . . . . . . . . . . . 137
- Binormaleneinheitsvektor . . . . . . . . . . . . . . . . . . . . . . . . . . 137
components
- Komponenten . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31, 65
components of the Christoffel symbol Komponenten des Christoffel-Symbols . . . . . . . . . . . . . . 142
components of the permutation tensor Komponenten des Permutationstensors . . . . . . . . . . . . . . . . 87
- kartesische Basisvektoren . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
composition
congruence transformation
Cartesian basis
- kartesische Basis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149
Cartesian components of a permutation tensor kartesische Komponenten des Permutationstensor . . . . . . 87
congruent
- kongruent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56, 63
Cartesian coordinates
continuum
- Kontinuum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
contravariant symbol
- kontravariantes -Symbol . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
Cauchys inequality
Cayley-Hamilton Theorem
- Cayley-Hamilton Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
characteristic equation
- charakteristische Gleichung . . . . . . . . . . . . . . . . . . . . . . . . . . 65
characteristic matrix
- charakteristische Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
characteristic polynomial
contravariant coordinates
- kontravariante Koordinaten, Koefzienten . . . . . . . . . 80, 84
contravariant metric coefcients kontravariante Metrikkoefzienten . . . . . . . . . . . . . . . . 82, 83
Christoffel symbol
- Christoffel-Symbol . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142
coordinates
- Koordinaten . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
cofactor
covariant symbol
- kovariantes -Symbol . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
column
- Spalte . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
column index
- Spaltenindex . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
column matrix
column vector
covariant coordinates
combination
- Kombination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9, 34, 54
covariant derivative
commutative
- kommutativ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
commutative matrix
- kommutative Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
cross product
commutative rule
- Kommutativgesetz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
commutative under matrix addition kommutativ bzgl. Matrizenaddition . . . . . . . . . . . . . . . . . . . 42
compatibility of vector and matrix norms Vertrglichkeit von Vektor- und Matrix-Norm . . . . . . . . . . 22
compatible
- vertrglich . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
curl
- Rotation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
curvature
- Krmmung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
curvature of a curve
curved surface
curvilinear coordinates
denite metric
denite norm
- denite Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
244
245
deformation energy
- Formnderungsenergie . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
dot product
- Punktprodukt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
deformation gradient
- Deformationsgradient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
dual space
- Dualraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36, 97
derivative of a scalar
derivative of a tensor
dummy index
- stummer Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
derivative of a vector
- Ableitung eines Vektors . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
derivative w.r.t. a scalar variable Ableitung nach einer skalaren Gre . . . . . . . . . . . . . . . . 133
dyadic product
eigenvalue
derivatives
- Ableitungen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
eigenvalue problem
eigenvalues
- Eigenwerte . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22, 56
determinant
- Determinante . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50, 65, 89
determinant expansion by minors Determinantenentwicklungssatz mit Unterdeterminanten51
eigenvector
eigenvector matrix
- Eigenvektormatrix, Modalmatrix . . . . . . . . . . . . . . . . . . . . . 70
determinant of a tensor
- Determinante eines Tensors . . . . . . . . . . . . . . . . . . . . . . . . . 112
determinant of the contravariant metric coefcients Determinante der kontravarianten Metrikkoefzienten . . 83
determinant of the contravariant metric coefcients Determinante der kovarianten Metrikkoefzienten. . . . . .83
determinant of the Jacobian matrix Determinante der Jacobimatrix . . . . . . . . . . . . . . . . . . . . . . 140
elastic
- elastisch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
elasticity tensor
- Elastizittstensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
elasticity theory
- Elastizittstheorie . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
elements
- Elemente . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
empty set
- leere Menge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
diagonal matrix
- Diagonalmatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41, 43
equilibrium conditions
- Gleichgewichtsbedingungen . . . . . . . . . . . . . . . . . . . . . . . . . 96
equilibrium conditon of moments Momentengleichgewichtsbedingung . . . . . . . . . . . . . . . . . 120
equilibrium system of external forces Gleichgewicht der ueren Krfte . . . . . . . . . . . . . . . . . . . . 96
- Krftegleichgewicht . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
dimension
- Dimension . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13, 14
Euclidean norm
- direkte Methode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
Euclidean space
- Euklidischer Raum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
direct product
- direktes Produkt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
Euclidean vector
- Hauptspannungsrichtungen . . . . . . . . . . . . . . . . . . . . . . . . . 120
deviator matrix
- Deviatormatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
direct method
discrete metric
- diskrete Metrik . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
distance
- Abstand . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
- Euklidische Matrixnorm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
even permutation
- gerade Permutation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
exact differential
distributive
- distributiv . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
distributive law
- Distributivgesetz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
eld
- Distributivgesetz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
eld
- Krper . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
nite
- endlich . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
- Finite-Element-Methode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
divergence theorem
- Divergenztheorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
domain
- Denitionsbereich . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
246
247
Frechet derivative
inner product
free indices
- freier Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
function
- Funktion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
fundamental tensor
- Fundamentaltensor. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .150
integers
- ganze Zahlen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
integral theorem
- Integralsatz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
intersection
- Schnittmenge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Gauss transformation
- Gausche Transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
Gausss theorem
- allgemeines Eigenwertproblem . . . . . . . . . . . . . . . . . . . . . . . 69
- allgemeines Permutationssymbol . . . . . . . . . . . . . . . . . . . . . 92
gradient
- Gradient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
invariance
- Invarianz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
invariant
invariant
- invariant . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
inverse
- Inverse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8, 34
inverse of a matrix
- inverse Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
inverse of a tensor
homeomorphic
- homomorph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30, 97
inverse relation
homeomorphism
- Homomorphismus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
inverse transformation
homogeneous
- homogen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
homogeneous linear equation system homogenes lineares Gleichungssystem . . . . . . . . . . . . . . . . 65
homogeneous norm
- homogene Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
inversion
- Umkehrung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
invertible
- invertierbar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
homomorphism
- Homomorphismus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
Hookes law
isomorphic
- isomorph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29, 35
- Hldersche Ungleichung . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
isomorphism
- Isomorphimus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
identities for scalar products of tensors Rechenregeln fr Skalarprodukte von Tensoren . . . . . . . 110
identities for tensor products
isotropic
- isotrop . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
isotropic tensor
iterative process
- Iterationsvorschrift . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
Jacobian
- Jacobi-Determinante . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
Kronecker delta
identity matrix
- Einheitsmatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41, 45
l-innity-norm, maximum-norm -
identity matrix
- Einheitsmatrix, Identitt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
identity tensor
- Einheitstensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
Maximumnorm, -Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
Laplace operator
- Laplace-Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
- Laplace-Operator eines Skalarfeldes . . . . . . . . . . . . . . . . . 150
image set
- Bildbereich . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
innitesimal
- innitesimal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
innitesimal tetrahedron
- innitesimaler Tetraeder . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
injective
- injektiv . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
- innerer Produktraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
line element
- Linienelement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
248
249
linear
- linear . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
linear algebra
- lineare Algebra . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
metric space
- metrischer Raum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
metric tensor of covariant coefcients Metriktensor mit kovarianten Koefzienten . . . . . . . . . . . 102
linear combination
linear dependence
- lineares Gleichungssytem . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
linear form
- Linearform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
- Momentengleichgewichtsbedingung . . . . . . . . . . . . . . . . . 120
linear independence
moving trihedron
linear manifold
- lineare Mannigfaltigkeit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
multiple roots
- Mehrfachnullstellen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
linear mapping
multiplicative identity
- multiplikationsneutrales Element . . . . . . . . . . . . . . . . . . . . . 45
linear operator
- linearer Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
multiplicative inverse
linear space
- linearer Raum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
linear subspace
- linearerUnterraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
linear transformation
- lineare Transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
- linearer Vektorraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
linearity
- Linearitt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32, 34
linearly dependent
linearly independent
lowering an index
main diagonal
mixed components
- gemischte Komponenten . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
mixed formulation of a second order tensor gemischte Formulierung eines Tensors zweiter Stufe . . . 99
n-tuple
- n-Tupel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
nabla operator
- Nabla-Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
natural basis
natural numbers
- natrliche Zahlen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
naturals
- natrliche Zahlen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
negative denite
- negativ denit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
Newtons relation
- Vietaschen Wurzelstze . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
- Hauptdiagonale . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41, 65
non-commutative
- nicht-kommutativ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
map
- Abbildung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
noncommutative
mapping
- Abbildung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
nonsingular
matrix
- Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
matrix calculus
- Matrizenalgebra . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
nonsymmetric
matrix multiplication
- Matrizenmultiplikation . . . . . . . . . . . . . . . . . . . . . . . . . . . 42, 54
nontrivial solution
matrix norm
- Matrix-Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21, 22
norm
- Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18, 65
norm of a tensor
normal basis
matrix transpose
- transponierte Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
maximum absolute column sum norm Spaltennorm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
maximum absolute row sum norm Zeilennorm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
maximum-norm
- Maximumsnorm, p-Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
mean value
- Mittelwert . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153
metric
- Metrik . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
metric coefcients
- Metrikkoefzienten . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
normal unit
- Normaleneinheitsvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
normal vector
normed space
- normierter Raum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
null mapping
- Nullabbildung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
odd permutation
- ungerade Permutation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
250
251
one
- Einselement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
- Hauptachsenproblem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
operation
- Operation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
principal axis
- Hauptachse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
operation addition
- Additionsoperation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
- Hauptspannungsrichtungen . . . . . . . . . . . . . . . . . . . . . . . . . 122
operation multiplication
- Multplikationsoperation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
principal stresses
- Hauptspannungen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
order of a matrix
product
- Produkt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
origin
- Ursprung, Nullelement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
orthogonal
- orthogonal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
orthogonal matrix
- orthogonalen Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
quadratic form
- Normquadrate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
- Heben eines Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
orthogonal tensor
orthogonal transformation
- orthogonale Transformation . . . . . . . . . . . . . . . . . . . . . . . . . . 57
raising an index
orthonormal basis
range
- Bildbereich . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
- Urbild . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
- Rang . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
outer product
- ueres Produkt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
range
overlined basis
rank
parallelepiped
- Parallelepiped . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
partial derivatives
- partielle Ableitungen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
partial derivatives of base vectors partielle Ableitungen von Basisvektoren . . . . . . . . . . . . . 145
rational numbers
- rationale Zahlen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Rayleigh quotient
- Rayleigh-Quotient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67, 68
real numbers
- reelle Zahlen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
rectangular matrix
- Rechteckmatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
- Rangabfall . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
permutation symbol
reduction of rank
permutation tensor
- Permutationstensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
permutations
- Permutationen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
point of origin
- Koordinatenursprung, -nullpunkt . . . . . . . . . . . . . . . . . . . . . 28
Poissons ratio
- Querkontraktionszahl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
polar decomposition
polynomial factorization
- Polynomzerlegung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
position vector
positive denite
positive metric
- positive Metrik . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
positive norm
- positive Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
post-multiplication
- Nachmultiplikation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
potential character
- Potentialeigenschaft. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .129
roots
- Nullstellen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
rotation matrix
- Drehmatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
rotation transformation
- Drehtransformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
rotator
- Rotor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116
row
- Zeile . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
row index
- Zeilenindex . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
row matrix
- Zeilenmatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40, 45
row vector
- Zeilenvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40, 45
- Skalarfeld . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
power series
- Potenzreihe . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
scalar eld
pre-multiplication
- Vormultiplikation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
scalar function
- Skalarfuntkion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
principal axes
- Hauptachsen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
scalar invariant
252
scalar multiplication
253
square
- quadratisch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
- quadratische Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
- multiplikationsneutrales Element . . . . . . . . . . . . . . . . . . . . . 10
square matrix
scalar product
Stokes theorem
strain tensor
stress state
- Spannungszustand . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
stress tensor
stress vector
- Spannungsvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
subscript index
- untenstehender Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
subset
- Untermenge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Schwarz inequality
summation convention
- Summenkonvention . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
second derivative
superscript index
- obenstehender Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
superset
- Obermenge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
supremum
- obere Schranke . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
section surface
- Schnittche . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
surface
- Oberche . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
semidenite
- semidenit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
surface element
- Oberchenelement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
Serret-Frenet equations
surface integral
- Oberchenintegral . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
set
- Menge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
surjective
- surjektiv . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
- Mengenlehre . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
symbols
- Symbole . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
shear stresses
- Schubspannungen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
symmetric
- symmetrisch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25, 41
similar
symmetric matrix
- symmetrische Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
similarity transformation
symmetric metric
- symmetrische Metrik . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
symmetric part
- symmetrischer Anteil . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
tangent unit
- Tangenteneinheitsvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
- Tangenteneinheitsvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
smmetric tensor
tangent vector
- Tangentenvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
space
- Raum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Taylor series
space curve
- Raumkurve . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
tensor
- Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
tensor axioms
- Axiome fr Tensoren . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
tensor eld
- Tensorfeld . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
set theory
span
- Hlle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
tensor product
- spezielles Eigenwertproblem . . . . . . . . . . . . . . . . . . . . . . . . . 65
spectral norm
- Spektralnorm, Hilbert-Norm . . . . . . . . . . . . . . . . . . . . . . . . . 22
tensor space
- Tensorraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
254
tensor with contravariant base vectors and covariant coordinates Tensor mit kontravarianten Basisvektoren und kovaranten
Koefzienten . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
tensor with covariant base vectors and contravariant coordinates Tensor mit kovarianten Basisvektoren und kontravaranten
Koefzienten . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
tensor-valued function of multiple variables tensorwertige Funktion mehrerer Vernderlicher . . . . . . 134
255
vector
vector eld
- Vektorfeld . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
vector function
- Vektorfunktion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
vector norm
- Vektor-Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18, 22
- Richtungsvektoren . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
vector of position
- Ortsvektoren . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
vector product
- Vektorprodukt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
vector space
- Vektorraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12, 49
vector-valued function
- vektorwertige Funktion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
vector-valued function of multiple variables vektorwertige Funktion mehrerer Vernderlicher . . . . . . 134
topology
- Topologie . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
torsion of a curve
total differential
trace of a matrix
trace of a tensor
transformation matrix
visual space
- Anschauungsraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
volume
- Volumen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
volume element
- Volumenelement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
volume integral
- Volumenintegral . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
volumetric matrix
- Kugelmatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
transformation relations
whole numbers
- natrliche Zahlen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Youngs modulus
- Elastizittsmodul . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
zero element
- Nullelement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
zero vector
- Nullvektor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
zeros
- Nullstellen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
- Transforamtionsformeln . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
transformation tensor
- Transformationstensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
transformed contravariant base vector transformierte kontravarianter Basisvektor . . . . . . . . . . . 103
transformed covariant base vector transformierte kovarianter Basisvektor . . . . . . . . . . . . . . . 103
transpose of a matrix
- transponierte Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
- transponiertes Matrizenprodukt . . . . . . . . . . . . . . . . . . . . . . 44
transpose of a tensor
triangle inequality
- Dreiecksungleichung . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16, 19
trivial solution
- triviale Lsung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
union
- Vereinigungsmenge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
unit matrix
- Einheitsmatrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
unitary space
- unitrer Raum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
- unitrer Vektorraum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
- bliches Skalarprodukt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
256
- L2-norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
n-Tupel
- n-tuple . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
Abbildung
- map . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Abbildung
- mapping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
- derivatives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
- distance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
additionsneutrales Element
Additionsoperation
- operation addition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
additiv
- additive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
hnlich, kogredient
- similar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55, 69
- outer product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
afner Vektor
- afne vector . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
afner Vektorraum
- visual space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
257
258
antisymmetrischer Anteil
- antisymmetric part . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
antisymmetrischer Anteil eines Tensors antisymmetric part of a tensor . . . . . . . . . . . . . . . . . . . . . . . 115
assoziativ
- associative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
assoziativ bzgl. Matrizenaddition associative under matrix addition . . . . . . . . . . . . . . . . . . . . . 42
Assoziativgesetz
- associative rule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Axiome fr Tensoren
- tensor axioms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
Basis
- basis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
259
Deviatormatrix
- deviator matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
Diagonalmatrix
differentielles Flchenelement
Basisvektoren
Dimension
- dimension . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13, 14
begleitendes Dreibein
direkte Methode
Betrag
- absolute value . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
direktes Produkt
- direct product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
diskrete Metrik
- discrete metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
bijektiv
- bijective . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
distributiv
- distributive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
Bildbereich
- image set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Distributivgesetz
- distributive law . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Bildbereich
- range . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Distributivgesetz
bilinear
- bilinear . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
Bilinearform
- bilinear form . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
Binormaleneinheitsvektor
Divergenztheorem
Binormaleneinheitsvektor
Drehmatrix
Drehtransformation
- rotation transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
Dreiecksungleichung
Cauchy-Spannungstensor
Cayley-Hamilton Theorem
- Cayley-Hamilton Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
charakteristische Gleichung
- characteristic equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
charakteristische Matrix
- characteristic matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
charakteristisches Polynom
Christoffel-Symbol
Dualraum
dyadisches Produkt
denite Metrik
- denite metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
denite Norm
- denite norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
Denitionsbereich
- domain . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Eigenwert
Deformationsgradient
Eigenwerte
- eigenvalues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22, 56
Determinante
Eigenwertproblem
260
einfacher Tensor vierter Stufe
261
gemischte Formulierung eines Tensors zweiter Stufe mixed formulation of a second order tensor . . . . . . . . . . . . 99
gemischte Komponenten
- mixed components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
gerade Permutation
- even permutation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
Einheitsmatrix
Einheitsmatrix
- unit matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
Einheitsmatrix, Identitt
- identity matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
Einheitstensor
Gesamtnorm
- absolute norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
Gleichgewicht der ueren Krfte equilibrium system of external forces . . . . . . . . . . . . . . . . . 96
Einselement
- one . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Gleichgewichtsbedingungen
- equilibrium conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
elastisch
- elastic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
Gradient
- gradient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
Elastizittsmodul
Elastizittstensor
Elastizittstheorie
Hauptachse
- principal axis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
Elemente
- elements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
endlich
- nite . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
Euklidische Matrixnorm
euklidische Norm
euklidische Vektoren
Euklidischer Raum
- Euclidean space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
euklidischer Vektorraum
Feld
- eld . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
Finite-Element-Methode
Flchenvektor
Formnderungsenergie
Frechet Ableitung
freier Index
- free indices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
Frenetsche Formeln
Fundamentaltensor
- fundamental tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
Fundamentaltensor dritter Stufe third order fundamental tensor . . . . . . . . . . . . . . . . . . . . . . 128
Funktion
- function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Hauptachsen
Hauptachsenproblem
Hauptdiagonale
Hauptspannungen
Hauptspannungsrichtungen
Hauptspannungsrichtungen
- raising an index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
homogen
- homogeneous . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
homogene Norm
- homogeneous norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
homogenes lineares Gleichungssystem homogeneous linear equation system . . . . . . . . . . . . . . . . . 65
Homomorphismus
- homomorphism . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
homomorph
- homeomorphic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30, 97
Homomorphismus
- homeomorphism . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
Hookesche Gesetz
Hldersche Ungleichung
Hlle
- span . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
innitesimal
- innitesimal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
innitesimaler Tetraeder
- innitesimal tetrahedron . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
ganze Zahlen
- integers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
injektiv
- injective . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Gaussscher Integralsatz
innerer Produktraum
Gausche Transformation
- Gauss transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
innerer Produktraum
gedrehtes Koordiantensystem
inneres Produkt
262
263
kommutative Matrix
- commutative matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
Integralnorm, L1-Norm
- L1-norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
Kommutativgesetz
- commutative rule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Integralsatz
komplexe Zahlen
- complex numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
invariant
- invariant . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
Invariante
Invarianz
- invariance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
Inverse
- inverse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8, 34
Komponenten
- components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31, 65
Komponenten des Christoffel-Symbols components of the Christoffel symbol . . . . . . . . . . . . . . . . 142
Komponenten des Permutationstensors components of the permutation tensor . . . . . . . . . . . . . . . . . 87
inverse Beziehung
Komposition
inverse Matrix
- inverse of a matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
inverse Transformation
inverser Tensor
kongruent
- congruent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56, 63
Kongruenztransformation, kontragrediente Transformation congruence transformation . . . . . . . . . . . . . . . . . . . . . . . 56, 63
konjugiert komplexe Eigenwerte complex conjugate eigenvalues . . . . . . . . . . . . . . . . . . . . . . 124
- invertible . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
isomorph
- isomorphic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29, 35
Isomorphimus
- isomorphism . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
isotrop
- isotropic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
isotroper Tensor
Iterationsvorschrift
- iterative process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
Jacobi-Determinante
- Jacobian . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
kartesische Basis
kartesische Basisvektoren
- Cartesian base vectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
kartesische Komponenten des Permutationstensor Cartesian components of a permutation tensor . . . . . . . . . 87
Kontinuum
- continuum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
- contravariant symbol . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
Koordinaten
- coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
Koordinatenursprung, -nullpunkt point of origin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
kovariante Ableitung
kovariante Basisvektoren
- covariant base vectors . . . . . . . . . . . . . . . . . . . . . . . . . . 80, 138
kovariante Basisvektoren der natrlichen Basis covariant base vectors of the natural basis . . . . . . . . . . . . 140
kovariante Koordinaten, Koefzienten covariant coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81, 84
kovariante Metrikkoefzienten - covariant metric coefcients . . . . . . . . . . . . . . . . . . . . . . 80, 83
kartesische Koordinaten
- Cartesian coordinates . . . . . . . . . . . . . . . . . . . . . . . 78, 82, 144
Kofaktor, algebraisches Komplement cofactor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
Kreuzprodukt
Kombination
Kronecker-Delta
- combination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9, 34, 54
kovariantes -Symbol
- covariant symbol . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
kommutativ
- commutative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
kommutativ bzgl. Matrizenaddition commutative under matrix addition . . . . . . . . . . . . . . . . . . . 42
krummlinige Koordinaten
- curvilinear coordinates. . . . . . . . . . . . . . . . . . . . . . . . .139, 144
krummliniges Koordinatensystem curvilinear coordinate system . . . . . . . . . . . . . . . . . . . . . . . 139
264
265
Krftegleichgewicht
Matrix-Norm
Krmmung
- curvature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
Matrizenalgebra
- matrix calculus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
Matrizenmultiplikation
Kugelmatrix
- volumetric matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
Maximumnorm, -Norm
Krper
- eld . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Laplace-Operator
- Laplace operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
Laplace-Operator eines Skalarfeldes laplacian of a scalar eld . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
Laplace-Operator eines Tensorfeldes laplacian of a tensor eld . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
Laplace-Operator eines Vektorfeldes laplacian of a vector eld . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
- l-innity-norm, maximum-norm . . . . . . . . . . . . . . . . . . . . . . 19
Maximumsnorm, p-Norm
- p-norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
Maximumsnorm, p-Norm
- maximum-norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
Mehrfachnullstellen
- multiple roots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
Menge
- set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
Mengenlehre
- set theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
Metrik
- metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
Metrikkoefzienten
- metric coefcients . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
Metriktensor mit kovarianten Koefzienten metric tensor of covariant coefcients . . . . . . . . . . . . . . . . 102
leere Menge
- empty set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
metrischer Raum
- metric space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
linear
- linear . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
linear abhngig
linear unabhngig
Mittelwert
- mean value . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153
Momentengleichgewichtsbedingung equilibrium conditon of moments . . . . . . . . . . . . . . . . . . . 120
Momentengleichgewichtsbedingung moment equilibrium conditon . . . . . . . . . . . . . . . . . . . . . . . 120
multiplikationsneutrales Element multiplicative identity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
multiplikationsneutrales Element scalar multiplication identity . . . . . . . . . . . . . . . . . . . . . . . . . 10
lineare Abbildung
lineare Abhngigkeit
lineare Algebra
- linear algebra . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
lineare Mannigfaltigkeit
- linear manifold . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
lineare Transformation
- linear transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
lineare Unabhngigkeit
linearer Operator
- linear operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
linearer Raum
Multplikationsoperation
- operation multiplication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
- linear space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
n-Tupel
- n-tuple . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
linearer Vektorraum
Nabla-Operator
linearerUnterraum
- linear subspace . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
Nachmultiplikation
- post-multiplication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
lineares Gleichungssytem
natrliche Basis
Linearform
- linear form . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
natrliche Zahlen
- natural numbers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .6
Linearitt
- linearity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32, 34
natrliche Zahlen
- naturals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Linearkombination
natrliche Zahlen
- whole numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
- negative denite . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
Linienelement
negativ denit
linker Cauchy-Strecktensor
Matrix
- matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
266
nicht kommutativ
- nontrivial solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
nicht-kommutativ
- non-commutative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
Norm
- norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18, 65
normale Basis
Normaleneinheitsvektor
Normaleneinheitsvektor
Normalenvektor
normierter Raum
Normquadrate
Nullabbildung
- null mapping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
Nullelement
- zero element . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Nullstellen
- roots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
267
partielle Ableitungen
- partial derivatives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
partielle Ableitungen von Basisvektoren partial derivatives of base vectors . . . . . . . . . . . . . . . . . . . . 145
Permutationen
- permutations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
Permutationssymbol
Permutationstensor
polare Zerlegung
Polynomzerlegung
- polynomial factorization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
positiv denit
positive Metrik
- positive metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
positive Norm
- positive norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
Potentialeigenschaft
Potenzreihe
- power series . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
Nullstellen
- zeros . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
Produkt
- product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Produkt von Tensoren zweiter Stufe second order tensor product . . . . . . . . . . . . . . . . . . . . . . . . . 105
Nullvektor
- zero vector . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Punktprodukt
obenstehender Index
quadratisch
- square . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
obere Schranke
- supremum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
quadratische Form
- dot product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
Oberche
- surface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
quadratische Matrix
- square matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Oberchenelement
Querkontraktionszahl
Oberchenintegral
Obermenge
- superset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Rang
- rank . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
Operation
- operation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
Rangabfall
- reduction of rank . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
rationale Zahlen
- rational numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
- order of a matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
orthogonal
- orthogonal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
orthogonale Transformation
- orthogonal transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
orthogonalen Matrix
- orthogonal matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
orthogonaler Tensor
orthonormale Basis
Ortsvektor
Raumkurve
Ortsvektoren
Parallelepiped
- parallelepiped . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
Rayleigh-Quotient
- Rayleigh quotient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67, 68
Rechenregeln fr Skalarprodukte von Tensoren identities for scalar products of tensors . . . . . . . . . . . . . . . 110
Raum
- space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Raum der quadratischen Matrizen space of square matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
268
- rectangular matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
rechter Cauchy-Strecktensor
reelle Zahlen
- real numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Richtungsvektoren
Riesz Abbildungssatz
Rotation
- curl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
Rotor
- rotator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116
scheifsymmetrische Matrix
- antisymmetric matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
schief- oder antisymmetrischer Anteil eines Tensors skew part of a tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
269
skalarwertige Vektorfunktion
Spalte
- column . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Spaltenindex
- column index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Spaltenmatrix
Spaltennorm
Spaltenvektor
Spannungstensor
Spannungsvektor
- stress vector . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
Spannungszustand
- stress state . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
Spatprodukt
Spektralnorm, Hilbert-Norm
- spectral norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
spezielles Eigenwertproblem
- trace of a matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
schiefsymmetrisch
- antisymmetric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
Schnittche
stummer Index
- dummy index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
- intersection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Summenkonvention
- summation convention . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
Schnittmenge
Schubspannungen
- shear stresses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
Schwarzsche oder Cauchy-Schwarzsche Ungleichung Cauchys inequality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
Summennorm, l1-Norm
- l1-norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
surjektiv
- surjective . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Symbole
- symbols . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
Schwarzsche Ungleichung
symmetrisch
- symmetric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25, 41
semidenit
- semidenite . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
symmetrische Matrix
- symmetric matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
- lowering an index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
symmetrische Metrik
- symmetric metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
skalare Invariante
skalare Multiplikation
Skalarfeld
symmetrischer Anteil
- symmetric part . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
symmetrischer Anteil eines Tensors symmetric part of a tensor . . . . . . . . . . . . . . . . . . . . . . . . . . 115
Skalarfuntkion
symmetrischer Tensor
Skalarprodukt
Tangenteneinheitsvektor
Tangenteneinheitsvektor
Tangentenvektor
Taylor-Reihe
skalarwertige Skalarfunktion
Tensor
- tensor. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
270
Tensor erster Stufe
271
berstrichene Basis
bliches Skalarprodukt
Umkehrung
- inversion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
ungerade Permutation
- odd permutation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
unitrer Raum
- unitary space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
unitrer Vektorraum
- unitary vector space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
unsymmetrisch, nicht symmetrisch nonsymmetric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
untenstehender Index
- subscript index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
Tensorfeld
Untermenge
- subset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Urbild
- range . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Ursprung, Nullelement
- origin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Vektor
Vektor-Norm
Tensorprodukt
- tensor product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105, 106
Tensorprodukt zweier Dyadenprodukte tensor product of two dyads . . . . . . . . . . . . . . . . . . . . . . . . . 106
Tensorraum
- tensor space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
tensorwertige Funktion mehrerer Vernderlicher tensor-valued function of multiple variables . . . . . . . . . . 134
Vektorfeld
tensorwertige Skalarfunktion
Vektorfunktion
tensorwertige Vektorfunktion
Vektorprodukt
- vector product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
Topologie
- topology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
Vektorraum
- vector space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12, 49
Vektorraum der linearen Abbildungen vector space of linear mappings . . . . . . . . . . . . . . . . . . . . . . 33
Transforamtionsformeln
- transformation relations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
Transformation der Basisvektoren transformation of base vectors . . . . . . . . . . . . . . . . . . . . . . 101
Transformation der Metrikkoefzienten transformation of the metric coefcients . . . . . . . . . . . . . . . 84
Transformationsmatrix
- transformation matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
Transformationstensor
- transformation tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
transformierte kontravarianter Basisvektor transformed contravariant base vector . . . . . . . . . . . . . . . . 103
transformierte kovarianter Basisvektor transformed covariant base vector . . . . . . . . . . . . . . . . . . . 103
transponierte Matrix
- matrix transpose . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
transponierte Matrix
- transpose of a matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
Transponierter Tensor
- transpose of a tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
transponiertes Matrizenprodukt transpose of a matrix product . . . . . . . . . . . . . . . . . . . . . . . . 44
triviale Lsung
- trivial solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
vektorwertige Funktion
- vector-valued function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
vektorwertige Funktion mehrerer Vernderlicher vector-valued function of multiple variables . . . . . . . . . . 134
vektorwertige Skalarfunktion
vektorwertige Vektorfunktion
Vereinigungsmenge
- union . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
vertrglich
- compatible . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
Vertrglichkeit von Vektor- und Matrix-Norm compatibility of vector and matrix norms . . . . . . . . . . . . . . 22
Verzerrungstensor, Dehnungstensor strain tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
Vietaschen Wurzelstze
- Newtons relation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
vollstndige Tensor zweiter Stufe complete second order tensor . . . . . . . . . . . . . . . . . . . . . . . . 99
vollstndiger Tensor dritter Stufe complete third order tensor . . . . . . . . . . . . . . . . . . . . . . . . . 129
TU Braunschweig, CSE Vector and Tensor Calculus 22. Oktober 2003
272
vollstndiges Differential
Volumen
- volume . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
Volumenelement
Volumenintegral
Vormultiplikation
- pre-multiplication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
Zeile
- row . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Zeilenindex
- row index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
Zeilenmatrix
Zeilennorm
Zeilenvektor
zweite Ableitung
Index
L1-norm, 19
L2-norm, 19
l1-norm, 19
l2-norm, Euclidian norm, 19
n-tuple, 28
p-norm, 19
absolute norm, 22
absolute value, 85
absolute value of a tensor, 111, 113
addition, 12, 13
additive, 32
additive identity, 10, 12, 42
additive inverse, 10, 12, 42
afne vector, 28
afne vector space, 28, 54
antisymmetric, 41
antisymmetric matrix, 41
antisymmetric part, 44
antisymmetric part of a tensor, 115
area vector, 152
associative, 42
associative rule, 10, 12
associative under matrix addition, 42
base vectors, 31, 87
basis, 31
basis of the vector space, 15
bijective, 8
bilinear, 25
bilinear form, 26
binormal unit, 137
binormal unit vector, 137
Cantor, 6
Cartesian base vectors, 88
273
Index
274
continuum, 96
contravariant symbol, 93
contravariant base vectors, 81, 139
contravariant base vectors of the natural basis, 141
contravariant coordinates, 80, 84
contravariant metric coefcients, 82, 83
coordinates, 31
covariant symbol, 92
covariant base vectors, 80, 138
covariant base vectors of the natural basis,
140
covariant coordinates, 81, 84
covariant derivative, 146, 149
covariant metric coefcients, 80, 83
cross product, 87, 90, 96
curl, 150
curvature, 135
curvature of a curve, 136
curved surface, 138
curvilinear coordinate system, 139
curvilinear coordinates, 139, 144
denite metric, 16
denite norm, 18
deformation energy, 129
deformation gradient, 118
derivative of a scalar, 133
derivative of a tensor, 134
derivative of a vector, 133
derivative w.r.t. a scalar variable, 133
derivatives, 133
derivatives of base vectors, 141, 145
determinant, 50, 65, 89
determinant expansion by minors, 51
determinant of a tensor, 112
determinant of the contravariant metric coefcients, 83
determinant of the Jacobian matrix, 140
deviator matrix, 46
deviator part of a tensor, 113
diagonal matrix, 41, 43
differential element of area, 139
dimension, 13, 14
direct method, 68
direct product, 94
directions of principal stress, 120
discrete metric, 17
distance, 17
distributive, 42
distributive law, 10, 13
distributive w.r.t. addition, 10
distributive w.r.t. scalar addition, 13
distributive w.r.t. vector addition, 13
divergence of a tensor eld, 147
divergence of a vector eld, 147
divergence theorem, 156
domain, 8
dot product, 85
dual space, 36, 97
dual vector space, 36
dummy index, 78
dyadic product, 9496
eigenvalue, 65, 120
eigenvalue problem, 22, 65, 122, 123
eigenvalues, 22, 56
eigenvector, 65, 120
eigenvector matrix, 70
Einstein, 78
elastic, 129
elasticity tensor, 129
elasticity theory, 129
elements, 6
empty set, 7
equilibrium conditions, 96
equilibrium conditon of moments, 120
equilibrium system of external forces, 96
equlibirum system of forces, 96
Euclidean norm, 22, 30, 85
Euclidean space, 17
Euclidean vector, 143
Euclidean vector space, 26, 29, 143
Euklidian matrix norm, 60
even permutation, 50
exact differential, 133, 135
Index
275
integers, 7
integral theorem, 156
intersection, 7
invariance, 57
invariant, 57, 120, 123, 148
inverse, 8, 34
inverse of a matrix, 48
inverse of a tensor, 115
inverse relation, 103
inverse transformation, 101, 103
inverse w.r.t. addition, 10, 12
inverse w.r.t. multiplication, 10
inversion, 48
invertible, 48
isomorphic, 29, 35
isomorphism, 35
isotropic, 129
isotropic tensor, 119
iterative process, 68
Jacobian, 140
Kronecker delta, 52, 79, 119
l-innity-norm, maximum-norm, 19
Laplace operator, 150
laplacian of a scalar eld, 150
laplacian of a tensor eld, 151
laplacian of a vector eld, 150
left-hand Cauchy strain tensor, 117
Leibnitz, 50
line element, 139
linear, 32
linear algebra, 3
linear combination, 15, 49, 71
linear dependence, 23, 30, 62
linear equation system, 48
linear form, 36
linear independence, 23, 30
linear manifold, 15
linear mapping, 32, 54, 97, 105, 121
linear operator, 32
linear space, 12
linear subspace, 15
Index
276
linear transformation, 32
linear vector space, 12
linearity, 32, 34
linearly dependent, 15, 23, 49
linearly independent, 15, 23, 48, 59, 66
lowering an index, 83
main diagonal, 41, 65
map, 8
mapping, 8
matrix, 40
matrix calculus, 28
matrix multiplication, 42, 54
matrix norm, 21, 22
matrix transpose, 41
maximum absolute column sum norm, 22
maximum absolute row sum norm, 22
maximum-norm, 20
mean value, 153
metric, 16
metric coefcients, 138
metric space, 17
metric tensor of covariant coefcients, 102
mixed components, 99
mixed formulation of a second order tensor,
99
moment equilibrium conditon, 120
moving trihedron, 137
multiple roots, 70
multiplicative identity, 45
multiplicative inverse, 10
n-tuple, 35
nabla operator, 146
natural basis, 140
natural numbers, 6, 7
naturals, 7
negative denite, 62
Newtons relation, 66
non empty set, 13
non-commutative, 45
noncommutative, 69, 106
nonsingular, 48, 59, 66
Index
277
Index
278
surface element, 152
surface integral, 152
surjective, 8
symbols, 6
symmetric, 25, 41
symmetric matrix, 41
symmetric metric, 16
symmetric part, 44
symmetric part of a tensor, 115
tangent unit, 135
tangent unit vector, 135
tangent vector, 135
Taylor series, 133, 153
tensor, 96
tensor axioms, 98
tensor eld, 143
tensor product, 105, 106
tensor product of two dyads, 106
tensor space, 94
tensor with contravariant base vectors and
covariant coordinates, 100
tensor with covariant base vectors and contravariant coordinates, 99
tensor-valued function of multiple variables,
134
tensor-valued scalar function, 133
tensor-valued vector function, 143
third order fundamental tensor, 128
third order tensor, 127
topology, 30
torsion of a curve, 137
total differential, 133
trace of a matrix, 43
trace of a tensor, 112
transformation matrix, 55
transformation of base vectors, 101
transformation of the metric coefcients, 84
transformation relations, 103
transformation tensor, 101
transformed contravariant base vector, 103
transformed covariant base vector, 103
transpose of a matrix, 41