Let (V,+,⋅) be a vector space. Let r, s be nonnegative integers. An (r,s)-tensor over V - let us call it T - is a multilinear map [1], [2], [3]
T:r timesV∗×…×V∗×s timesV×…×V→IR,
where V∗ is the dual vector space of V - the space of covectors.
Before we explain what multilinearity is let's give an example.
Let's say T is a (1,1)-tensor. Then T:V∗×V→IR, that is T takes a covector and a vector and multilinearity means that for two covectors ϕ,ψ∈V∗,
T(ϕ+ψ,v)=T(ϕ,v)+T(ψ,v),
and
T(λϕ,v)=λT(ϕ,v),
for all v∈V, and it is also linear in the second argument, that is,
T(ϕ,v+w)=T(ϕ,v)+T(ϕ,w),
and
T(ϕ,λv)=λT(ϕ,v).
Using these properties twice we can see that
T(ϕ+ψ,v+w)=T(ϕ,v)+T(ϕ,w)+T(ψ,v)+T(ψ,w).
Example: A (0,2)-tensor. Let us denote the set of polynomials with real coefficients is denoted by P. Define the map
g:(p,q)↦g(p,q):=∫01p(x)q(x)dx.
Then this is a (0,2)-tensor,
g:P×P→IR,
and multilinearity is easy to check.
It becomes evident that an inner product is a (0,2)-tensor. More precisely, an inner product is a bilinear symmetric positive definite operator V×V→IR.
Example: Covectors as tensors. It is easy to see that covectors, ϕ:V→IR are (0,1)-tensors. Trivially.
Example: Vectors as tensors. Given a vector v∈V we can consider the mapping
Tv:V∗∋ϕ↦Tv(ϕ)=ϕ(v)∈IR.
This makes Tv into a (1,0)-tensor.
Example: linear maps as tensors. We can see a linear map ϕ∈Hom(V,V) as a (1,1)-tensor
Let T be an (r,s)-tensor over a finite-dimensional vector space V and let {ei}i be a basis for V. Let the dual basis, for V∗, be {ϵi}i. The two bases have the same cardinality. Then define the (r+s)dimV many numbers
These are the components of T with respect to the given basis.
This allows us to conceptualise a tensor as a multidimensional (multi-indexed) array. But maybe we shouldn’t… This is as bad as treating vectors as sequences of numbers, or matrices as “tables” instead of elements of a vector space and linear maps respectively.
Let's see how exactly this works via an example [4]. Indeed, consider an (2,3)-tensor of the tensor space T32=V⊗2⊗(V∗)⊗3 , which can be seen as a multilinear map
T:V∗×V∗×V×V×V→IR.
Yes, we have two V∗ and three V!
Take a basis (ei)i of V and a basis (θj)j of V∗. Let us start with an example using a pure tensor of the form ei1⊗ei2⊗ei3⊗θj1⊗θj2, for some indices i1,i2,i3 ,and j1,j2. This can be seen as a map V∗×V∗×V×V×V→IR, which is defined as
Here we construct a huge vector space and apply a quotient to enforce the axioms we expect tensors and tensor products to have. This huge space is a space of functions sometimes referred to as the "formal product" [5]. See also this video [7].
We will define the tensor product of two vector spaces. Let V,W be two vector spaces. We define a vector space V∗W, which we will call the formal product of V and W, as the linear space that has V×W as a Hamel basis. This space is also known as the free vector space, V∗W=Free(V×W).
To make this more concrete, we can identify V∗W by the space of functions φ:V×W→IR with finite support. Representatives (and a basis) for this set are the functions
δv,w(x,y)={1,0, if (x,y)=(v,w) otherwise
Indeed, every function f:V×W→IR with finite support (a function of V∗W) can be written as a finite combination of such δ functions and each δ function is identified by a pair (v,w)∈V×W.
Note that V×W is a vector space when equipped with the natural operations of function addition and scalar multiplication.
We consider the natural embedding, δ, of V×W into V∗W, which is naturally defined as
We define the tensor product of V with W as
V⊗W=M0V∗W.
This is called the tensor space of V with W and its elements are called tensors.
This is the space we were looking for. Here's what we mean: we have already defined the mapping δ:V×W→V∗W. We also define the canonical embedding π:V∗W→V⊗M0. We then define the tensor product of v∈V and w∈W as
v⊗w=(π∘δ)(v,w)=δv,w+M0.
It is a mapping ⊗:V×W→V⊗W and we can see that it is bilinear.
Here's how I understand the universal property: Suppose we know that we have a function f:V×W→?, which is bilinear. We can always think of the mysterious space ? as the tensor space V⊗W[5].
Let's look at Figure 1.
Figure 1. Universal property of tensor product.
There is a unique linear function f~:V⊗W→? such that
f(v,w)=f~(v⊗w).
Let us underline that f~ is linear! This makes ⊗ a prototype bilinear function as any other bilinear function is a linear map of precisely this one.
Let V and W be finite dimensional. We will show that
Dimension of tensor
dim(V⊗W)=dimVdimW.
Proof 1. To that end we use the fact that the dual vector space has the same dimension as the original vector space. That is, the dimension of V⊗W is the dimension of (V⊗W)∗.
The space (V⊗W)∗=Hom(V⊗W,IR) is the space of bilinear maps V×W→IR.
Suppose V has the basis {e1V,…,enVV} and W has the basis {e1W,…,enWW}.
To form a basis for the space of bilinear maps f:V×W→IR we need to identify every such function with a sequence of scalars. We have
f(u,v)=f(i=1∑nVaiVeiV,i=1∑nWaiWeiW).
From the bilinearity of f we have
f(u,v)=i=1∑nVj=1∑nWaiVajWf(eiV,ejW).
The right hand side is a bilinear function and the coefficients (aiV,ajW) suggest that the dimension if nVnW. □
Proof. This is due to [22]. We can write any T∈V⊗W as
T=k=1∑nak⊗bk,
for ak∈V and bk∈W and some finite n. If we take bases {vi}i=1nV and {wi}i=1nW we can write
We need first to define the function space F(S) as in Kostrikin[2].
Let S be any set. We define the set F(S)—we can denote it also as Funct(S,IR)—as the set of all functions from S to IR. If f∈F(S), then f is a function f:S→IR and f(s) denotes the value at s∈S.
On F(S) we define addition and scalar multiplication in a pointwise manner: For f,g∈F(S) and c∈IR,
(f+g)(s)=(cf)(s)=f(s)+g(s),cf(s).
This makes F(S) into a vector space. Note that S is not necessarily a vector space.
If S={s1,…,sn} is a finite set, F(S) can be identified with IRn. After all, for every f∈F(S) all you need to know is f(s1),…,f(sn).
Every element s∈S is associated with the delta function
δs(σ)={1,0, if σ=s, otherwise
The function δs:S→{0,1} is called Kronecker's delta.
If S is finite, then every f∈S can be written as
f=s∈S∑asδs.
Let S1 and S2 be finite sets and let F(S1) and F(S2) be the corresponding function spaces. Then, there is a canonical identity of the form
Has δs1,s2 as basisF(Finite set of pairsS1×S2)=F(S1)⊗F(S2),
which associates each function δs1,s2 with δs1⊗δs2.
If f1∈F(S1) and f2∈F(S2) then using the standard bases of F(S1) and F(S2)
As Kostrikin notes [2], the spaces (L1⊗L2)⊗L3 and L1⊗(L2⊗L3) do not coincide — they are spaces of different type. They, however, can be found to be isomorphic via a canonical isomorphism.
We will start by studying the relationship between L1⊗L2⊗L3 and (L1⊗L2)⊗L3.
The mapping L1×L2∋(l1,l2)↦l1⊗l2∈L1⊗L2 is bilinear, so the mapping
Note that v∗⊗w is a pure tensor. A general tensor has the form
i∑vi∗⊗wi,
and
Φ(i∑vi∗⊗wi)(v)=i∑vi∗(v)wi.
This is a linear map. Linearity is easy to see. See also this video [6].
Here is the statement:
First isomorphism result: Hom(V,W)≅V∗⊗W.
In the finite-dimensional case, we have that Hom(V,W)≅V∗⊗W holds with isomorphism ϕ⊗w↦(v↦ϕ(v)w).
Proof.[10] The proposed map, ϕ⊗w↦(v↦ϕ(v)w), is a linear map V∗⊗W→Hom(V,W). We will construct its inverse (and prove that it is the inverse). Let (ei)i be a basis for V. Let (ei∗)i be the naturally induced basis of the dual vector space, V∗. We define the mapping
g:Hom(V,W)∋ϕ↦i∑ei∗⊗ϕ(ei)∈V∗⊗W.
We claim that g is the inverse of Φ which we already defined to be the map V∗⊗W→Hom(V,W) given by Φ(v∗⊗w)(v)=v∗(v)w. Indeed, we see that
Φ(g(ϕ))=====Φ(i=1∑nei∗⊗ϕ(ei))i=1∑nΦ(ei∗⊗ϕ(ei))i=1∑nei∗(⋅)ϕ(ei)ϕ⎝⎜⎛idei∗(⋅)ei⎠⎟⎞ϕ.Definition of gΦ is linearϕ is linear
It is Hom(V,W)∗≅Hom(W,V).{\rm Hom}(V, W)^* {}\cong{} {\rm Hom}(W, V).Hom(V,W)∗≅Hom(W,V).
Some additional consequences of this are:
Hom(Hom(V,W),U)≅Hom(V,W)∗⊗U≅(V∗⊗W)∗⊗U≅V⊗W∗⊗U.{\rm Hom}({\rm Hom}(V,W),U) \cong {\rm Hom}(V,W)^*\otimes U \cong (V^*\otimes W)^*\otimes U \cong V\otimes W^* \otimes U.Hom(Hom(V,W),U)≅Hom(V,W)∗⊗U≅(V∗⊗W)∗⊗U≅V⊗W∗⊗U.
U⊗V⊗W≅Hom(U∗,V)⊗W≅Hom(Hom(V,U∗),W)U\otimes V \otimes W \cong {\rm Hom}(U^*, V)\otimes W \cong {\rm Hom}({\rm Hom}(V,U^*),W)U⊗V⊗W≅Hom(U∗,V)⊗W≅Hom(Hom(V,U∗),W)
V⊗V∗≅Hom(V∗,V)≅Hom(V,V)V\otimes V^* \cong {\rm Hom}(V^*, V) \cong {\rm Hom}(V,V)V⊗V∗≅Hom(V∗,V)≅Hom(V,V), where if VVV is finite dimensional, V∗≅VV^*\cong VV∗≅V, so we see that V∗⊗V≅End(V)V^*\otimes V \cong {\rm End}(V)V∗⊗V≅End(V)
Question: How can we describe linear maps from I Rm×n{\rm I\!R}^{m\times n}IRm×n to I Rp×q{\rm I\!R}^{p\times q}IRp×q with sensors?
We are talking about the space Hom(I Rm×n,I Rp×q){\rm Hom}({\rm I\!R}^{m\times n}, {\rm I\!R}^{p\times q})Hom(IRm×n,IRp×q), but every matrix AAA can be identified by the linear map x↦Axx\mapsto Axx↦Ax, so I Rm×n≅Hom(I Rn,I Rm){\rm I\!R}^{m\times n}\cong {\rm Hom}({\rm I\!R}^{n}, {\rm I\!R}^{m})IRm×n≅Hom(IRn,IRm), so
Such objects can be seen as multilinear maps I Rn∗×I Rm×I Rq×I Rp∗→I R{\rm I\!R}^{n*} \times {\rm I\!R}^{m} \times {\rm I\!R}^{q} \times {\rm I\!R}^{p*} \to {\rm I\!R}IRn∗×IRm×IRq×IRp∗→IR , that is,
We can define V⊗WV\otimes WV⊗W to be the a space accompanied by a bilinear operation ⊗:V×W→V⊗W\otimes: V\times W \to V \otimes W⊗:V×W→V⊗W that has the universal property in the following sense: [5]
Universal property. If f:V×W→Zf: V\times W \to Zf:V×W→Z is a bilinear function, then there is a unique linear function f~:V⊗Z\tilde{f}: V\otimes Zf~:V⊗Z such that f=f~∘⊗f = \tilde{f} \circ \otimesf=f~∘⊗.
The universal property says that every bilinear map f(u,v)f(u, v)f(u,v) on a vector space VVV is a linear function f~\tilde{f}f~ of the tensor product, f~(u⊗v)\tilde{f}(u\otimes v)f~(u⊗v)[6].
See more about the uniqueness of tensor product via the universality-based definition as well as[8].
This is based on[9]. We can identify vectors u=(u0,…,un−1)∈I Rnu=(u_0, \ldots, u_{n-1})\in{\rm I\!R}^nu=(u0,…,un−1)∈IRn and v∈I Rmv\in{\rm I\!R}^mv∈IRm as polynomials over I R{\rm I\!R}IR of the form
We know that a pair of vectors lives in the Cartesian product space I Rn×I Rm{\rm I\!R}^n\times {\rm I\!R}^mIRn×IRm. A pair of polynomials lives in a space with basis
Lots of books on tensors for physicists are in [11]. In Linear Algebra Done Wrong[12] there is an extensive chapter on the matter. It would be interesting to read about tensor norms [13]. These lecture notes [14][15][16] seem worth studying too. These [18] lectures notes on multilinear algebra look good, but are more theoretical and look a bit category-theoretical. A must-read book is [19] and these lectures notes for MIT [20]. A book that seems to explain things in an accessible way, yet rigorously is [21].
Prof Macauley on YouTube, Advanced Linear Algebra, Lecture 3.7: Tensors, accessed on 9 November 2023; see also his slides, Lecture 3.7: Tensors, lecture slides on tensors, some nice diagrams and insights.
Answer on MSE on why Hom(V,W){\rm Hom}(V, W)Hom(V,W) is the same thing as V∗⊗WV^*\otimes WV∗⊗W, accessed on 9 November 2023
Lots of books on tensors, tensor calculus, and applications to physics can be found on this GitHub repo, accessed on 9 November 2023
MIT Multilinear algebra lecture notes/book, multilinear algebra, introduction (dual spaces, quotients), tensors, the pullback operation, alternating tensors, the space Λk(V∗)\Lambda^k(V^*)Λk(V∗), the wedge product, the interior product, orientations, and more 👍 (must read)
JR Ruiz-Tolosa and E Castillo, From vectors to tensors, Springer, Universitext, 2005.
Elias Erdtman, Carl Jonsson, Tensor Rank, Applied Mathematics, Linkopings Universite, 2012