Linear Vector Space

Let us denote the linear vector space formed by these automorphism, or to be more concrete, by regular matrices representing these automorphisms, as Autg.

From: Exterior Analysis , 2013

Basic Review

George A. Articolo , in Partial Differential Equations & Boundary Value Problems with Maple (Second Edition), 2009

0.2 Preparation for Linear Algebra

A linear vector space consists of a set of vectors or functions and the standard operations of addition, subtraction, and scalar multiplication. In solving ordinary and partial differential equations, we assume the solution space to behave like an ordinary linear vector space. A primary concern is whether or not we have enough of the correct vectors needed to span the solution space completely. We now investigate these notions as they apply directly to two-dimensional vector spaces and differential equations.

We use the simple example of the very familiar two-dimensional Euclidean vector space R2; this is the familiar (x, y) plane. The two standard vectors in the (x, y) plane are traditionally denoted as i and j. The vector i is a unit vector along the x-axis, and the vector j is a unit vector along the y-axis. Any point in the (x, y) plane can be reached by some linear combination, or superposition, of the two standard vectors i and j. We say the vectors "span" the space. The fact that only two vectors are needed to span the two-dimensional space R2 is not coincidental; three vectors would be redundant. One reason for this has to do with the fact that the two vectors i and j are "linearly independent"—that is, one cannot be written as a multiple of the other. The other reason has to do with the fact that in an n-dimensional Euclidean space, the minimum number of vectors needed to span the space is n.

A more formal mathematical definition of linear independence between two vectors or functions v1 and v2 reads as "The two vectors v1 and v2 are linearly independent if and only if the only solution to the linear equation

c 1 υ 1 + c 2 υ 2 = 0

is that both c1 and c2 are zero." Otherwise, the vectors are said to be linearly dependent.

In the simple case of the two-dimensional (x, y) space R2, linear independence can be geometrically understood to mean that the two vectors do not lie along the same direction (noncolinear). In fact, any set of two noncolinear vectors could also span the vector space of the (x, y) plane. There are an infinite number of sets of vectors that will do the job. One common connection between all sets, however, is that all the sets can be shown to be linearly dependent; that is, all the sets can be shown to be reducible to linear combinations of the standard i and j vectors.

For example, the two vector sets

S 1 = { i , j }

are both linearly independent sets of vectors that span the two-dimensional (x, y) space. Note that the vectors within each set are linearly independent, but the vectors between sets are linearly dependent.

A set of vectors S = {v1, v2, v3, …, vn} that are linearly independent and that span the space is called a set of "basis" vectors for that particular vector space. Thus, for the two-dimensional Euclidean space R2, the vectors i and j form a basis, and for the three-dimensional Euclidean space R3, vectors i, j, and k form a basis. The number of vectors in a basis is called the "dimension" of the vector space.

A set of basis vectors is fundamental to a particular vector space because any vector in that space can then be written as a unique superposition of those basis vectors. These concepts are important to us when we consider the solution space of both ordinary and partial differential equations. Another important concept in linear algebra is that of the inner product of two vectors in that particular vector space.

For the Euclidean space R3, if we let u and v be two different vectors in this space with components

u = [ u 1 , u 2 , u 3 ]

then the inner product of these two vectors is given as

i p ( u , υ ) = u 1 υ 1 + u 2 υ 2 + u 3 υ 3

Thus, the inner product is the sum of the product of the components of the two vectors. The inner product is sometimes also referred to as the "dot product."

If we take the square root of an inner product of a vector with itself, then we are evaluating the length of the vector, commonly called the "norm."

n o r m ( u ) = i p ( u , u )

Different vector spaces have different inner products. For example, we consider the vector space C[a, b] of all functions that are continuous over the finite closed interval [a, b]. Let f(x) and g(x) be two different vectors in this space. The inner product of these two vectors over the interval, with respect to the weight function w(x), is defined as the definite integral:

i p ( f , g ) = a b f ( x ) g ( x ) w ( x ) d x

From the basic definition of a definite integral, we see the inner product to be an (infinite) sum of the product of the components of the two vectors.

Similarly, in the space of continuous functions, if we take the square root of the inner product of a vector with itself, then we evaluate the length or norm of the vector to be

n o r m ( f ) = a b f ( x ) 2 w ( x ) d x

As an example, consider the two functions f(x) = sin(x) and g(x) = cos(x) over the finite closed interval [0, π] with a weight function w(x) = 1. The length or norm of f(x) is the definite integral

n o r m ( f ) = 0 π sin ( x ) 2 d x

which evaluates to

n o r m ( f ) = π 2

Similarly, for g(x) the norm is the definite integral

n o r m ( g ) = 0 π cos ( x ) 2 d x

which evaluates to

n o r m ( g ) = π 2

If we evaluate the inner product of the two functions f(x) and g(x), we get the definite integral

i p ( f , g ) = 0 π cos ( x ) sin ( x ) d x

which evaluates to

i p ( f , g ) = 0

If the inner product between two vectors is zero, we say the two vectors are "orthogonal" to each other. Orthogonal vectors can also be shown to be linearly independent.

If we divide a vector by its length or norm, then we "normalize" the vector. For the preceding f(x) and g(x), the corresponding normalized vectors are

f ( x ) = 2 π sin ( x )

A set that consists of vectors that are both normal and orthogonal is said to be an "orthonormal" set. For orthonormal sets, the inner product of two vectors in the set gives the value 1 if the vectors are alike or the value 0 if the vectors are not alike.

Two vectors φn(x) and φm(x), which are indexed by the positive integers n and m, are orthonormal with respect to the weight function w(x) over the interval [a, b] if the following relation holds:

a b φ n ( x ) φ m ( x ) w ( x ) d x = δ ( n , m )

Here, δ(n, m) is the familiar Kronecker delta function whose value is 0 if n ≠ m and is 1 if n = m.

Orthonormal sets play a big role in the development of solutions to partial differential equations.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780123747327000032

Exterior Algebra

Erdoğan S. Şuhubi , in Exterior Analysis, 2013

1.3 Multilinear Functionals

Let (U 1, U 2, …,Uk ) be ordered k -tuple of linear vector spaces defined over the same field of scalars F . Let us consider a scalar-valued function T : U 1 × U 2 × × U k F on the Cartesian product of these vector spaces. If the function, T u 1 u 2 u k F , where u (α) ∈ U α , α = 1, 2,…, k, is linear in each one of its arguments, that is, if the following relations

(1.3.1) T , u i + υ i , = T u i + T υ i T , α u i , = α T u i , α F

are satisfied for all 1   ik, then the function T is called a multilinear functional (or a k-linear functional ). In finite-dimensional vector spaces whose dimensions and bases are n 1,.,nk and {e i (α)}   U α , i  =   1,   …, n α ,α  =   1,   …, k, we can then write u α = i = 1 n α u α i e i α , without having recourse to the summation convention. Multilinearity then leads to the following value of the functional at vectors u (1)  U 1, u (2)  U 2,   …, u (k)  U k

(1.3.2) T u 1 u 2 u k = i 1 = 1 n 1 i 2 = 1 n 2 i k = 1 n k t i 1 i 2 i k u 1 i 1 u 2 i 2 u k i k

Where n 1 × n 2 × ... × n k number of scalar t i 1 i 2 i k are defined by

(1.3.3) t i 1 i 2 i k = T e i 1 1 e i 2 2 e i k k F

We thus conclude that the set of scalars t i 1 i 2 i k completely determines the action of a k -linear functional on any set of k number of vectors u (1)  U 1, u (2)  U 2,   …, u (k)  U k . We can thus say that they unambiguously characterise a multilinear functional.

Let us now suppose that U 1 = U 2 =⋯ = Uk = U (n). The value of a multilinear functional T : U k F on vectors u (1),u (2), ..., u(k)U can now be found from (1.3.2) and (1.3.3) as follows

(1.3.4) T u 1 u 2 u k = t i 1 i 2 i k u 1 i 1 u 2 i 2 u k i k , t i 1 i 2 i k = T e i 1 e i 2 e ik , 1 i 1 , i 2 , , i k n

where we experience no difficulty in resorting to the summation convention because the range of all indices is the same now, from 1 to n. In this case, we can introduce a more advantageous representation of a multilinear functional as an operator. To this end, we shall first introduce the tensor product of two vector spaces.

Let U and V be two linear vector spaces defined on the same field of scalars F . As is well known, the Cartesian product U × V of these spaces is formed by ordered pairs (u,υ), where uU and υ ∈ V. There is initially no algebraic structure on this product set. However, by making use of known operations on vector spaces U and V, we may define appropriate operations on the set U × V so that it may be equipped with a structure of a linear vector space. The resulting vector space will be called the tensor product of spaces U and V and will be denoted by W  = U  V. Let us choose operations of vector addition and scalar multiplication on W in such a way that tensor product of vectors u  υ  U  V has to satisfy the following bilinearity conditions:

(i).

u    (υ 1  + υ 2)   = u  υ 1  υ 2,

(ii).

(u 1  + u 2)   υ  = u 1  υ  + u 2  υ,

(iii).

αu υ = u αυ = α u υ , α F .

Let us note that the same symbol + in the foregoing expressions represents,in fact, different addition operations in three different vector spaces U, V and W. We can thus write

u 1 + u 2 υ 1 + υ 2 = u 1 υ 1 + u 1 υ 2 + u 2 υ 1 + u 2 υ 2 .

The space W is then defined as the collection of all finite sums i u i υ i where ui U and υi V . If we consider finite-dimensional vector spacesU (m)and V (n)with respective bases {ei } and {fj }, a vector w  W is evidently expressible as ω  = ω ij e i   f j . Hence, W is an mn-dimensionalvector space with a basis {e i   f j }. The tensor product can evidently be extendedon Cartesian products of arbitrary number of vector spaces.Let us now consider the n-dimensional dual space U of an n-dimensionalvector space U. It is quite clear that an element, or a vector, of thetensor product ⊗ k U* can now be represented by

(1.3.5) T = t i 1 i 2 ... i k f i 1 f i 2 ... f i k

where{fi } is the reciprocal basis in U * corresponding to the basis {e i } in U. We define the value of the element T on an ordered k-tuple of vectors (u (1),u (2),   …,u (k))   U k as

T u 1 u k = t i 1 i k u 1 j 1 u k jk f i 1 e j 1 f i k e jk

In view of (1.2.7), we then find that

T u 1 u 2 u k = t i 1 i 2 i k u 1 i 1 u 2 i 2 u k i k .

We immediately see that the above relation leads to (1.3.4) 2 for vectors e i 1 , e i 2 , , e i k . Hence (1.3.5) does in fact play the part of k -linear functionalon Uk and the tensor product ⊗ k U* is the vector space in which such k-linear functionals inhabit. We say that the elements of this vector space are k-covariant tensors and the number k is known as the order of the tensor.The scalar coefficients t i 1 i 2 i k are then called the components of such a tensor with respect to bases. f i 1 f i k It is easily observed that thetensorproduct . f i 1 f i k of basis vectors constitutes a basis for the space ⊗ k U*. Indeed the value of the zero element in ⊗ k U*

t i 1 i 2 i k f i 1 f i 2 f i k = 0

on vectors e j 1 , e j 2 , , e j k U vanishes naturally so that one obtains

t i 1 i 2 i k f i 1 e j 1 f i 2 e j 2 f i k e j k = t j 1 j 2 j k = 0

for all coefficients. Hence, the dimension of this vector space is nk . Obviously,the sum of two tensors of the same kind and multiplication of a tensorby a scalar are again the following tensors of the same kind:

T 1 + T 2 = t i 1 i 2 i k 1 + t i 1 i 2 i k 2 f i 1 f i 2 f i k α T = α t i 1 i 2 i k f i 1 f i 2 f i k .

This is of course a direct consequence of ⊗ k U* being a linear vector space.

We can now naturally define the tensorial product of a k-covariant tensorand an l-covariant tensor by

T 1 T 2 = t i 1 i k 1 t j 1 j i 2 f i 1 f i k f j 1 f jl .

The result is obviously a (k + l)-covariant tensor.

Let us now change the basis {ei }in the vector space U to another basis {e i '} as in (1.2.11). We know that the reciprocal basis {fi }in the dual space U* changes to a reciprocal basis {f'i }through the relations (1.2.14). Consequently,the same tensor T is represented with respect to two different basesas follows

T = t j 1 j 2 j k f j 1 f j 2 f j k = t i 1 i 2 i k ' f ' i 1 f ' i 2 f ' ik = t j 1 j 2 j k b i 1 j 1 b i 2 j 2 b ik jk f ' i 1 f ' i 2 f ' i k

from which we immediately deduce that the following rule of transformation between components of a k-covariant tensor must be valid:

(1.3.6) t i 1 i 2 i k ' = b i 1 j 1 b i 2 j 2 b i k j k t j 1 j 2 j k .

In a similar fashion we may define a multilinear (k-linear) functionalon the dual space U* of a vector space. Such a functional T : U * k F assigns a scalar number T f 1 f 2 f k F to an ordered k-tuple of linear functionals (f (1),f (2),   …,f (k))     (U*) k and obeys the rules

T , f i + g i , = T ( , f i , ) + T ( , g i , ) T , α f i , = α T ( , f i , ) , α F .

By resorting to the reciprocal basis {f i }   U* corresponding to the basis {ei } ∈ U we can of course write f m = α i m f i , α i m F , 1 m k and we obtain

(1.3.7) T f 1 f 2 f k = t i 1 i 2 i k α i 1 1 α i 2 2 α i k k , t i 1 i 2 i k = T f i 1 f i 2 f i k .

The ensemble of scalar numbers t i 1 i 2 i k , 1 i 1 , i 2 , , i k n entirely determines the action of a multilinear functional T on (U  *) k . Let us nowdefine an element in the tensor product ⊗ k U by

T = t i 1 i 2 i k e i 1 e i 2 e i k .

T is called a k-contravariant tensor . It is evident that the linearly independentselements e i 1 e i 2 e i k constitute a basis for the vector space ⊗ k U. n k number of scalars t i 1 i 2 i k are said to be components of this tensorwith respect to bases e i 1 e i k . Let us define the value of the tensor T on k linear functionals f (1),f (2),…,f (k) by the relation

T f 1 f 2 f k = t i 1 i 2 i k f 1 e i 1 f 2 e i 2 f k e i k .

In view of (1.2.6) we find that

T f 1 , f 2 , f k = t i 1 i 2 i k α i 1 1 α i 2 2 i k k .

It is clear that the product of a k-contravariant tensor and an l-contravarianttensor is a (k + l)-contravariant tensor. We now consider a change of basisin the vector space U. We then obtain

T = t j 1 j 2 j k e j 1 e j 2 e j k = t ' i 1 i 2 i k e i 1 ' e i 2 ' e i k ' = t j 1 j 2 j k a j 1 i 1 a j 2 i 2 a j k i k e i 1 ' e i 2 ' e i k '

from which we deduce the following rule of transformation for componentsof a contravariant tensor

(1.3.8) t ' i 1 i 2 i k = a j 1 i 1 a j 2 i 2 a j k i k t j 1 j 2 j k .

We can also easily define tensors of mixed type. A k-contravariant and l-covariant mixed tensor is an element of the vector space ⊗ k U  l U  * and can be written in the form

T = t j 1 j 2 j l i 1 i 2 i k e i 1 e i 2 e i k f j 1 f j 2 f j l , t j 1 j 2 j l i 1 i 2 i k = T f i 1 f i 2 f i k e j 1 e j 2 e j l , 1 i 1 , i 2 , , i k n , 1 j 1 , j 2 , , j l n .

The value of this tensor on linear functionals f (1), f (2),   ⋯, f (k)  U  * and Vectors u (1), u (2),   ⋯, u (l)  U is given by

T f 1 f k u 1 u l = t j 1 j 2 j l i 1 i 2 i k α i 1 1 α i 2 2 α i k k u 1 j 1 u 2 j 2 u l j l .

It is quite obvious that we do not have to select the ordering in the tensorproducts in the foregoing way. We may, of course, consider a differentordering such as U  U  *   U  *   U  U  *     ⋯. The indices of componentsof this type of a tensor occupy accordingly proper upper and lower positions.It is evident that different ordering of spaces in the tensor product willgive rise to different types of tensors of the same order.

If, in a mixed tensor of order k + l, we remove the tensor product betweenthe functional f j m and the vector e i n , then the relation f j m e i n = δ i n j m between reciprocal basis vectors reduces the order of the tensor. We thusobtain a (k − 1)-contravariant and (l − 1)-covariant tensor, in other words, atensor of order k + l − 2 defined by the relation

T c = t j 1 i n j l i 1 i n i k e i 1 e i n 1 e i n + 1 e i k f j 1 f j m 1 f j m + 1 f j l .

This operation is called a contraction . The components of the contractedtensor are given as follows:

c t j 1 j m 1 j m + 1 j l i 1 i n 1 i n + 1 i k = t j 1 j m 1 i j m + 1 j l i 1 i n 1 i i n + 1 i k .

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780124159020500013

Fast Translations: Basic Theory and O(p3) Methods

NAIL A. GUMEROV , RAMANI DURAISWAMI , in Fast Multipole Methods for the Helmholtz Equation in Three Dimensions, 2004

7.1.1.2 Linear operators

Operators are defined as mappings from one linear vector space to another (both spaces can be the same). If we have a function ψ ∈ B(Ω) and can correspond it to some other function ψ′ ∈ B(Ω′), then such a mapping is an operator A acting from the space B(Ω) to the space B(Ω′). This can be written as

(7.1.1) ψ = A [ ψ ] , A : B ( Ω ) B ( Ω ) , ψ B ( Ω ) , ψ B ( Ω )

The operator is called linear if it satisfies the following property:

(7.1.2) A [ α ψ 1 + β ψ 2 ] = α A [ ψ 1 ] + β A [ ψ 2 ] , α , β .

Important examples of linear operators are as follows.

Differential operators, D. An example is ∂/∂z which performs partial differentiation with respect to z (we also used ∂ z = k −1∂/∂z and sometimes to stress that this operator acts in functional space, we use the notation Dz = ∂ z ). Differential operators include ∇, ∇2, or differentiation in a direction s, s·∇. These operators map B(Ω) → B(Ω), so that the domain of definition does not change and we have

(7.1.3) ψ = D [ ψ ] , D : B ( Ω ) B ( Ω ) .

Rotation operators, Rot. These operators are generated by rotation of the basis vectors with matrix Q, so we can write this operator as Rot(Q). If Ω is the interior or exterior of a sphere, then rotation does not change the domain of definition, and we have

(7.1.4) ψ = R o t ( Q ) [ ψ ] , ψ ( r ) = ψ ( r ^ ) , r ^ = Q r , R o t ( Q ) : B ( Ω ) B ( Ω ) , Ω : r a .

Translation operators, T. These operators arise due to a shift in the function argument by a translation vector t, so we can write them as T(t). The translation operator shifts the domain of definition of the transformed function ψ′(r) = ψ(r +t) (indeed if the function ψ(r) has only a singularity at r = -t, then ψ′(r) has the same singularity at r = 0). In general, we have

(7.1.5) ψ = T ( t ) [ ψ ] , ψ ( r ) = ψ ( r ) , r = r + t , T ( t ) : B ( Ω ) B ( Ω ) ,

where Ω′ is obtained by the shift of the domain Ω.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780080443713500119

Lie Groups: General Theory

R. Gilmore , in Encyclopedia of Mathematical Physics, 2006

Solvable Algebra

If g is a Lie algebra, the linear vector space obtained by taking all possible commutators of the operators in g is called the "derived" algebra: [ g , g ] = g ( 1 ) g . If g ( 1 ) = g , there is no point in continuing this process. If g ( 1 ) g , it is useful to define g = g ( 0 ) and to continue this process by defining g ( 2 ) as the derived algebra of g ( 1 ) : g ( 2 ) = [ g ( 1 ) , g ( 1 ) ] . We can continue in this way, defining g ( n + 1 ) as the algebra derived from g ( n ) . Ultimately (for finite-dimensional Lie algebras), either g ( n + 1 ) = 0 or g ( n + 1 ) = g ( n ) for some n. If the former case occurs,

g = g ( 0 ) g ( 1 ) g ( 2 ) g ( n ) g ( n + 1 ) = 0

the Lie algebra g ( 0 ) is called solvable. Each algebra g ( i ) is an invariant subalgebra of g ( j ) , i > j .

Example

The Lie algebra spanned by the boson number, creation, annihilation, and identity operators is solvable. The series of derived algebras has dimensions 4, 3, 1, 0.

g ( 0 ) g ( 1 ) g ( 2 ) g ( 3 )
a a
a a
a a
I I I

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B0125126662004259

Electromagnetic Theory and Optics

Avijit Lahiri , in Basic Optics, 2016

Digression: tensors and tensor fields

For a given r and given t, a vector such as E(r, t ) is an element of a real three-dimensional linear vector space which we denote as R ( 3 ) . A tensor of rank 2 is then an element of a nine-dimensional vector space T that includes the direct product R ( 3 ) × R ( 3 ) and, in addition, contains all possible linear combinations of direct products of pairs of vectors. If n ^ 1 , n ^ 2 , n ^ 3 constitute an orthonormal basis in R ( 3 ) , then an orthonormal basis in T will be made up of the objects n ^ i n ^ j ( i , j = 1 , 2 , 3 ) , and a tensor of rank 2 can be expressed as a linear combination of the form i , j C i j n ^ i n ^ j . Thus, with reference to this basis, the tensor under consideration is completely described by the 3 × 3 matrix with elements C ij . The matrix (and also the tensor) is termed 'symmetric' if C ij = C ji (i, j = 1, 2, 3). The matrix is said to be positive definite if all its eigenvalues are positive. Now consider any of the above field vectors (say, E(r, t)) at a given time instant but at all possible points r. This means a vector associated with every point in some specified region in space. The set of all these vectors is termed a vector field in the region under consideration. The vector field is, moreover, time dependent since the field vectors depend, in general, on t. Similarly, one can have a tensor field such as the permittivity tensor [ϵ] or the permeability tensor [μ] in an inhomogeneous anisotropic medium in which the electric and magnetic material properties vary from point to point in addition to being direction dependent. While these can, in general, even be time-dependent tensor fields, we will, in this book, consider media with time-independent properties alone.

Thus, in terms of the Cartesian components, relations (1.1f) and (1.1g) can be written as

(1.2a) D i = j ϵ i j E j ,

(1.2b) B i = j μ i j H j .

As mentioned above, the electric permittivity and magnetic permeability tensors ([ϵ], [μ]) reduce, in the case of an isotropic medium, to scalars (corresponding to constant multiples of the identity matrix) and the above relations simplify to

(1.3a) D = ϵ E , D i = ϵ E i ( i = 1 , 2 , 3 ) ,

(1.3b) B = μ H , B i = μ H i ( i = 1 , 2 , 3 ) .

It is not unusual for an optically anisotropic medium, with a permittivity tensor [ϵ], to be characterized by a scalar permeability μ (approximately μ 0, the permeability of free space). In this book I use the SI system of units, in which the permittivity and permeability of free space are, respectively, ϵ 0 = 8.85 × 10−12 C2 N−1 m−2 and μ 0 = 4π × 10−7 NA−2.

In general, for linear media with time-independent properties, the following situations may be encountered: (1) isotropic homogeneous media, for which ϵ and μ are scalar constants independent of r; (2) isotropic inhomogeneous media for which ϵ and μ are scalars but vary from point to point; (3) anisotropic homogeneous media where [ϵ] and [μ] are tensors independent of the position vector r; and (4) anisotropic inhomogeneous media in which [ϵ] and [μ] are tensor fields. As mentioned above, in most situations relating to optics one can, for simplicity, assume [μ] to be a scalar constant, μμ 0.

However, in reality, the relation between E and D is of a more complex nature (that between B and H may, in principle, be similarly complex), even for a linear, homogeneous, isotropic medium with time-independent properties, than is apparent from Eq. (1.3a) since ϵ is, in general, a frequency-dependent object. A time-dependent field vector can be analyzed into its Fourier components, each component corresponding to some specific angular frequency ω. A relation such as Eq. (1.3a) can be used only in situations where this frequency dependence of the electric (and also magnetic) properties of the medium under consideration can be ignored (ie, when dispersion effects are not important). In this book we will generally assume the media are nondispersive, taking into account dispersion effects only in certain specific contexts (see Section 1.17).

One more constitutive equation holds for a conducting medium:

(1.4) j = [ σ ] E ,

where, in general, the conductivity [σ] is once again a second-rank symmetric tensor which, for numerous situations of practical relevance, reduces to a scalar. The conductivity may also be frequency dependent, as will be discussed briefly in Section 1.17.2.7.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128053577000010

Tensor Fields on Manifolds

Erdoğan S. Şuhubi , in Exterior Analysis, 2013

4.2 Cotangent Bundle

We consider an m-dimensional smooth manifold M and the tangent space Tp (M) at a point pM . As is well known, the dual of the tangent space is a linear vector space formed by all linear functionals on the tangent space [see p. 11]. We denote this m-dimensional dual space by T p (M) and we also call it the cotangent space at the point p. When we choose the natural basis of the tangent space at the point p as the vectors {∂/∂xi : i = 1, … , m} generated by the local coordinates in the chart containing the point p, we have seen on p. 125 that reciprocal basis vectors in the dual space are given by linear functionals as differentials {dxi : i = 1, , m} so that the following relations

(4.2.1) d x i x j = d x i , x j = δ j i

are satisfied. Hence, at a point pM, a vector VTp (M) and a linear functional ω  T p *(M) can be expressed as

(4.2.2) V = υ i x i , ω = ω i d x i , υ i , ω i R

The value of the functional ω on the vector V at p then happens to be

(4.2.3) ω V = ω V = ω i d x i , υ j x j = ω i υ j δ j i = ω i υ i R

We shall call elements of the dual space T p *(M) as 1-forms at the point p. Next, we define the set

(4.2.4) T * M = p M T p * M = p ω : p M , ω T p * M .

By repeating exactly our approach in Sec. 2.8, we see that T*(M) can be endowed with a differentiable structure making it a 2m-dimensional smooth manifold which will be called henceforth as the cotangent bundle . The local coordinates of T*(M) are evidently given by {x 1,…, xm , ω 1, … , ωm }. A section of the bundle T*(M) as we have already done in p. 130 characterises this time a 1-form field on the smooth manifold M. In terms of local coordinates in the relevant chart, this field is of course expressible as follows

(4.2.5) ω p = ω i x d x i T * M , x = φ p

Different charts containing the point p gives rise to a coordinate transformation given by invertible functions yi = yi (xj ). When we write the 1-form ω in different local coordinates, the relation

ω p = ω j d x j = ω i d y i = ω i y i x j d x j

leads to the following relations between components of ω in two different coordinate systems

ω j = ω i y i x j or ω i = x j y i ω j

Because of this transformation rule, the elements of the cotangent bundle are usually called covariant vector or covector fields . We have already seen that the transformation rule between components of vectors in two different charts in the tangent bundle are given by [see (2.6.9)]

υ i = y i x j υ j

That is the reason why we call vectors in the tangent bundle as contravariant vector fields.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780124159020500049

Representations for Morphological Image Operators and Analogies with Linear Operators

Petros Maragos , in Advances in Imaging and Electron Physics, 2013

2.2.1 Algebraic Definitions for Linear Operators

A mapping L : X Y between two linear spaces over the same scalar field is called a linear operator if it preserves addition and scalar multiplication. This is equivalent to

L ( a 1 x 1 + + a n x n ) = a 1 L ( x 1 ) + + a n L ( x n )

for all x 1 , , x n X , all scalars a 1 , , a n and all finite n. An operator that does not satisfy the above is called nonlinear. The null space and range of L are defined as

Null ( L ) { x X : L ( x ) = 0 } , Ran ( L ) { L ( x ) : x X } .

The null space is a linear subspace of X, whereas the range is a linear subspace of Y. For any linear operator L : X Y between two linear spaces, the dimensions of its null and range spaces are related as follows to the dimension of the domain space:

dim(Nul ( L )) + dim(Ran ( L )) = dim ( Dom ( L ) ) .

Two linear spaces X and Y over the same scalar field are called isomorphic if there exists an invertible linear operator L : X Y . Such a mapping L is called an isomorphism between the two linear spaces. The inverse mapping L 1 : Y X is also a linear operator. It can be shown that a linear operator L is an isomorphism iff Nul ( L ) = { 0 } .

All linear spaces over the same field are isomorphic iff they have the same dimension. Hence, all real (resp. complex) finite-dimensional linear spaces are isomorphic to R n (resp. C n ) for some n. Thus, finite-dimensional linear spaces are essentially linear vector spaces, if by "vector" § we agree to mean a finite tuple of scalars.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780124077027000024

Preliminaries

Dan C. Marinescu , Gabriela M. Marinescu , in Classical and Quantum Information, 2012

1.4 POSTULATES OF QUANTUM MECHANICS

A model of a physical system is an abstraction based on correspondence rules that relate the entities manipulated by the model to the physical objects or systems in the real world. Once such rules are established, we can operate only with the abstractions according to a set of transformation rules. To ensure the usefulness of the model and its ability to describe physical reality, we have to validate the model and compare its prediction with the physical reality. To ensure expressiveness, the ability of the model to describe the physical system, the correspondence and the transformation rules must be kept as simple as possible, but, at the same time, complete—in other words, capable of capturing the relevant properties of the physical system and of its dynamics, its evolution in time.

Distinguishability and system dynamics require the model to abstract the concepts of observable and of state of the physical object. An observable is a property of the system state that can be revealed as a result of some physical transformation. The state at time t is a synthetic characterization of the object that could be revealed by the measurement of relevant observables at time t.

The model must also abstract the concept of measurement; it should describe the relation between the state of the object before and after the measurement and specify how to interpret the results of a measurement, how to map the range of possible results to abstractions. In the physical world, we often have to deal with a collection of physical objects. If A, B, C, … are the abstractions of the objects a, b, c, …, respectively, we need another transformation rule to specify how to construct {A, B, C, …}, the abstraction corresponding to the collection {a, b, c, …}. Last, but not least, we need transformation rules to describe the system dynamics, the evolution of the system in time.

Quantum mechanics is a model of the physical world at all scales; it describes more accurately than classical physics systems at the atomic and subatomic scale. This model allows us to abstract our knowledge of a quantum system, to describe the state of single and composite quantum systems, the effect of a measurement on the system's state, and the dynamics of quantum systems. A quantum state summarizes our knowledge about a quantum system at a given moment in time, it allows us to describe what we know, as well as, what we do not know, about the system. An impressive number of experiments have produced results consistent with the prediction of quantum mechanics and so far there is no experimental evidence to disprove it; thus, we shall use this model to study the properties of quantum information.

The correspondence and transformation rules are captured by the postulates of quantum mechanics (Figure 1.2). We find it useful to expand the traditional three postulates of quantum mechanics, the state postulate, the dynamics postulate, and the measurement postulate, to emphasize some aspects important for quantum information processing:

FIGURE 1.2. The postulates of quantum mechanics.

1.

A quantum system, Q, is described in an n-dimensional Hilbert space, Hn , where n is finite. The Hilbert space Hn is a linear vector space over the field of complex numbers with an inner product. The dimension, n, of the Hilbert space is equal to the maximum number of reliably distinguishable states the system Q can be in.

2.

A state |ψ〉 of the quantum system Q corresponds to a direction (or ray) in Hn . In Section 1.11, we shall see that the most general representation of a quantum state is any density operator over an n-dimensional Hilbert space with n finite. The density operator is Hermitian, has non-negative eigenvalues, and has a trace equal to unity.

3.

When the internal conditions and the environment of a quantum system are completely specified and no measurements are performed on the system, the system's evolution is described by a unitary transformation in Hn defined by the Hamiltonian operator. A unitary transformation U is linear and preserves the inner product. The spontaneous evolution of an unobserved quantum system with the density matrix, ρ, is

ρ U ρ U ,

with U, the adjoint of U.
4.

Given two independently prepared quantum systems, Q described in Hn and S described in Hn , the bipartite system consisting of both Q and S is described, in a Hilbert space, Hn ⊗ Hm, the tensor product of the two Hilbert spaces.

5.

A measurement of the quantum system Q in the state |ψ described in Hn corresponds to a resolution of Hn to orthogonal subspaces, {H j }, and a projection of the system's state to these subspaces, {Pj}, such that the sum of the projections is P j = 1. The measurement produces the result, j, with the probability

Pr ob ( j ) = | P j | ψ | 2 .

The state after the measurement is

| φ = P j | ψ | P j | ψ | = P j | ψ Pr ob ( j ) .

Manipulation of coherent quantum states is at the heart of quantum computing and quantum communication. A quantum computation involves a single entity and consists of unitary transformations of the quantum state. Quantum communication involves multiple entities and involves the transmission of quantum states over noisy communication channels.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780123838742000011

Convex Functions, Partial Orderings, and Statistical Applications

In Mathematics in Science and Engineering, 1992

1.28 Remarks

(a) We shall say that f′ is increasing on U if for x, yU we have

( f ( x ) f ( y ) ) ( x y ) 0 ,

and that f′ is strictly increasing on U if this inequality is strict for all xy.

(b) A function f: UM (UL; L, M are normed linear vector spaces) is Fréchet differentiable at x 0 (x 0U) if there exists a linear transformation T: LM such that

lim x x 0 f ( x ) f ( x 0 ) T ( x x 0 ) x x 0 = 0 ,

which is equivalent to

f ( x ) = f ( x 0 ) + T ( x x 0 ) + o ( x x 0 )

as xx 0. The linear transformation T is called the Fréchet derivative and is denoted by f′(x 0).

(c) A similar derivative of the Fréchet derivative is called the second Fréchet derivative. This derivative is a symmetric bilinear transformation defined on L x L, i.e., f k, h (x) = f h, k (x) (h, kL). Note that if f: U is continuously differentiable on the open convex set UL and f″(x) exists throughout U, then for any x, x 0U there is an s ∈ (0, 1) such that

(1.27) f ( x ) = f ( x 0 ) + f h ' ( x 0 ) + 1 2 f h , h " ( x 0 + s h ) ,

where h = x − x 0.

(d) A symmetric bilinear transformation B(h, k) defined on L x L is positive (nonnegative) definite if for every hL (h ≠ 0), we have

B ( h , h ) > 0 ( B ( h , h ) 0 ) .

(e) The following definition is also valid: A (continuous real-valued) function f is operator convex on (λ, v) if fa + βb) ≤ αf(a) + βf(b) for positive reals α, β such that α + β = 1 and operators a, b with their spectra in (λ, v). (See Davis, 1957, for a brief survey of operator functions and Ando, 1978, for further comments on classes of operator functions.)

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/S0076539208628131

Mathematical preliminaries

J.E. Akin , in Finite Element Analysis with Error Estimators, 2005

2.1 Introduction

The earliest forms of finite element analysis were based on physical intuition with little recourse to higher mathematics. As the range of applications expanded, for example to the theory of plates and shells, some physical approaches failed and some succeeded. The use of higher mathematics such as variational calculus explained why the successful methods worked. At the same time the mathematicians were attracted by this new field of study. In the last few years the mathematical theory of finite element analysis has grown quite large. Since the state of the art now depends heavily on error estimators and error indicators it is necessary for an engineer to be aware of some basic mathematical topics of finite element analysis. We will consider load vectors and solution vectors, and residuals of various weak forms. All of these require us to define some method to 'measure' these entities. For the above linear vectors with discrete coefficients, V T = [V 1 V 2Vn ], we might want to use a measure like the root mean square, RMS:

R M S 2 = 1 n i = 1 n V i 2 = 1 2 V T V

which we will come to call a norm of the linear vector space. Other quantities vary with spatial position and appear in integrals over the solution domain and/or its boundaries. We will introduce various other norms to measure these integral quantities.

The finite element method always involves integrals so it is useful to review some integral identities such as Gauss' Theorem (Divergence Theorem):

Ω u d Ω = Γ u n d Γ = Γ u n d Γ

which is expressed in Cartesian tensor form as

Ω u i , i d Ω = Γ u i n i d Γ

where there is an implied summation over subscripts that occurs an even number of times and a comma denotes partial differentiation with respect to the directions that follow it.

That is, ( ), i = ∂( )/x i . The above theorem can be generalized to a tensor with any number of subscripts:

Ω A i j k q , r d Ω = Γ A i j k q n r d Γ .

We will often have need for one of the Green's Theorems:

Ω ( A B + A 2 B ) d Ω = Γ A B n d Γ

And

Γ ( A 2 B B 2 A ) d Ω = Γ ( A B B A ) n d Γ

which in Cartesian tensor form are

Ω ( A , i B , i + A B , i i ) d Ω = Γ A B , i n i d Γ

and

Ω ( A B , i i B A , i i ) d Ω = Γ ( A B , i B A , i ) n i d Γ .

We need these relations to derive the Galerkin weak form statements and to manipulate the associated error estimators. Usually, we are interested in removing the highest derivative term in an integral and use the second from last equation in the form

(2.1) Ω A B , i i d g Ω = Γ A B , i n , i d Γ Ω A , i B , i d Ω .

In one-dimensional applications this process is called integration by parts:

a b p d q = p q | a b a b q d p .

Error estimator proofs utilize inequalities like the Schwarz inequality

(2.2) | a b | | a | | b |

and the triangle inequality

(2.3) | a + b | | a | + | b | .

Finite element error estimates often use the Minkowski inequality

(2.4) [ i = 1 n | x i ± y i | p ] 1 / p [ i = 1 n | x i | p ] 1 / p + [ i = 1 n | y i | p ] 1 / p , 1 < p < ,

and the corresponding integral inequality

(2.5) [ Ω | x ± y | p d Ω ] 1 / p [ Ω | x | p d Ω ] 1 / 2 + [ Ω | y | p d Ω ] 1 / p , 1 < p < ,

We begin the preliminary concepts by introducing linear spaces. These are a collection of objects for which the operations of addition and scalar multiplication are defined in a simple and logical fashion.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780750667227500333