Although all applications of dot product were presented within the field of real numbers ℝ so far, this section can be applied to arbitrary scalar field, including complex numbers ℂ.
Duality
The dot product of two numerical vectors (i.e., vectors from 𝔽n, where 𝔽 is a field of either real numbers ℝ or complex numbers ℂ) is a single number that provides information
about the relationship between two vectors. The linearity of dot product tells us that the scalar product of two numerical vectors can be used for definition of linear forms/functionals. The reverse statement (see section on dual spaces in Part 3) is always true: every linear functional on a finite dimensional vector space is generated by a corresponding dot product.
Example 12:
In calculus, the gradient of a scalar function is a dual vector. It takes a vector (representing a direction) and returns the rate of change of the scalar function in that direction.
In physics, dual vectors appear in various contexts, such as in describing fields (like electromagnetic fields) and in quantum mechanics (using bra-ket notation).
■
End of Example 12
At the beginning of twentieth century, it was discovered that the dot product is needed for definition of dual spaces (see section in Part3). Although definition of dot product is symmetrical, it is convenient to separate vectors identifying one of them as a vector, called contravariant vector or ket-vector. When ordered basis in a finite dimensional vector space is chosen, [e₁, e₂, … , en], every vector can be uniquely expressed as linear combination of the basis vectors:
whose components are identified with superscripts. The upper indices are not exponents but are indices of coordinates, identifying the coefficients of basis vectors.
It is customary to omit the sum symbol in this expression adopting the Einstein
summation convention.
This convention is routinely used in Einstein’s general theory of
relativity which suffers from a proliferation of indices but it also facilitates a simplified
notation and more effective computations in many other contexts.
Using the dual basis [e¹, e², … , en], a vector from the dual space can be written as
This vector y from the dual space is called the covariant vector or covector or dual vector or bra-vector.
Then using the duality property \( \displaystyle \quad \mathbf{e}^j \mathbf{e}_i = \delta^i_j , \quad \) where
δij is the Kronecker's delta,
we can write the dot product of bra-vector y with ket-vector x in concise form:
defines a linear functional on any n dimensional vector space independently on what field is used (ℂ or ℝ).
Then one of the multipliers, say x in the linear combination \eqref{EqDot.2}, is called a vector, but another counterpart, y is known as a covector. Such treatment of vectors in the dot product breaks their "equal rights." In many practical problems, these vectors, x and y, are indeed taken from different spaces, although sometimes look the same.
In geometry, to distinguish these two partners in Eq.\eqref{EqDot.2}, the vector x is called contravariant vector, and the covector y is referred to as covariant vector. In order to decide between these partners who is who, it is common to use superscript for coordinates of contravariant vector, x = [ x¹, x², x³], and subscript for covariant vectors, y = [y₁, y₂, y₃]. In physics, covariant vectors are also called bra-vectors, while contravariant vectors are known as ket-vectors.
Duality in matrix/vector multiplication
Although duality is covered in Part 3, we introduce some of its elements here in order to show a different point of view in the context of matrix/vector multiplication.
So far, we have been using a naïve definition of a vector as an ordered (or indexed) finite sequence of scalars; it is usually represented in parenthesis as an n-tuple (x1, x2, … , xn), an element of Cartesian product 𝔽n = 𝔽 × 𝔽 × ⋯ × 𝔽 of n copies of the field of scalars 𝔽. Later, we will extend this definition to include abstract objects; however, for us at this moment, it is sufficient to work with n-tuples. Recall that for scalars we use either rational numbers ℚ or real numbers ℝ or complex numbers ℂ.
Usually, an n-tuple (x1, x2, … , xn), written in parenthesis and a row vector [x1, x2, … , xn], written in brackets, look as the same object to human eyes. One of the pedagogical virtues of any software package is its requirement to pay close attention to the types of objects used in specific contexts. In particular, the computer algebra system Mathematica treats these two versions of a vector differently because
where 𝔽m×n denotes the space of m × n matrices with coefficients from field 𝔽.
Hence, Mathematica considers bracket notation as a matrix with one row.
x = {1, 2}; y = {{1, 2}};
x == y
False
Mathematica considers single curly bracket notation as a list of entries, which is treated as a column vector for convenience. On the other hand, a double curly bracket notation is considered as a list of lists and it is represented as a matrix with one row.
It is a matrix with one row and one column. In order to extract this entry from 1×1 matrix, mathematicians consider this expression as a linear combination. It also can be defined by dot product:
when all entries are real numbers. You need to wait till Part 5 for dot product extension to complex numbers (it is inner product). Although dot product formula above can be used for complex numbers as well, it does not define a metric (distance) in complex vector spaces. Physicists usually use Dirac's notation (also known as bra-ket notation), which is designated as <a | x> for an alternative to linear combination or dot product.
Every numeric vector a = (𝑎1, 𝑎2, … , 𝑎n) ∈ 𝔽n determines
the corresponding dual vector (known as a covector) a* that defines a linear functional a* : 𝔽n ⇾ 𝔽 via the simple rule
When vector
a = (𝑎1, 𝑎2, … , 𝑎n) is considered as a covector, it is usually called a bra-vector and is denoted by <a∣, while x = (x1, x2, … , xn) ∈ ℂn is called a ket-vector or just a vector and denoted by ∣x>. Then the linear combination \( \displaystyle a_1 x_1 + a_2 x_2 + \cdots + a_n x_n \) (which is dot product for real numbers) defines a linear functional (or linear form), which we denote by a*.
Hence, every n-tuple a = (𝑎1, 𝑎2, … , 𝑎n) defines a covector (or linear functional). In case of real numbers, action of a covector on a vector is just dot product.
Example 1:
The linear functional on ℝ4 given by
\[
f\left( x_1 , x_2 , x_3 , x_4 \right) = 2\,x_1 - 3\, x_2 + x_3 - 5\, x_4
\]
is dual to the numeric vector a = (2, −3, 1, −5). This vector generates a linear functional on ℝ4 by means of dot product
\[
\langle {\bf a} \mid {\bf x} \rangle = {\bf a} \bullet {\bf x} = 2\,x_1 - 3\, x_2 + x_3 - 5\, x_4 ,
\]
for arbitrary vector x = (x₁, x₂, x₃, x₄) ∈ ℝ4. This functional, written in bra-ket notation, can be redefined in standard mathematical form as
\[
{\bf a}^{\ast} \, : \ \mathbb{R} \mapsto \mathbb{R} , \qquad\mbox{with} \qquad
{\bf a}^{\ast} ({\bf x}) = \langle {\bf a} \mid {\bf x} \rangle = {\bf a} \bullet {\bf x} .
\]
On the other hand, the numeric vector a = (2, −3, 1) ∈ ℝ³ is dual to the linear functional (known as a covector) given by
\[
{\bf a}^{\ast} \left( {\bf x} \right) = \left\langle {\bf a} \mid {\bf x} \right\rangle = {\bf a} \bullet {\bf x} = 2\, x_1 -3\, x_2 + x_3 , \quad {\bf x} \in \mathbb{R}^3 .
\]
In particular, actions of the covector 〈a | on some bra-vectors from ℝ³ are given below:
\begin{align*}
{\bf a}^{\ast} (1, 2, 3) &= {\bf a} \bullet \left( 1, 2, 3 \right) = 2\cdot 1 -3\cdot 2 + 3 = -1,
\\
{\bf a}^{\ast} (-1, 2, -5) &= {\bf a} \bullet \left( -1, 2, -5 \right) = 2\cdot (-1) -3 \cdot 2 -5 = -13 .
\end{align*}
End of Example 1
In Linear Algebra, the word duality traditionally refers to the dual
roles of vector and covector associated with a given numeric vector, as
in definition above. The row/column duality just described is one of many
beautiful dualities we highlight
in this tutorial. We will see many other examples of duality, including
image/inverse image, geometric/numeric, and projection/span—problems that
seem difficult from one viewpoint often turn out to be much easier
from the dual viewpoint.
Each of the two perspectives illuminates the
other in a beautiful and useful way, and we shall make good use of this
principle many times.
We can regard an m × n matrix, for instance, as a list of
m (row) vectors in ℝ1×n, or as a list of n (column) vectors in ℝm×1, and
the difference is not superficial, as we shall come to see.
Every m × n matrix
where "T" stands for transposition---a dual operation that transfers rows into columns and vice versa.
Since in the following exposition we utilize matrix multiplication from left, we are forced to consider vectors as columns. Then
In other words, A x is a linear combination of column vectors of matrix A with weights taken from coordinates of vector x.
Alternatively, using dual covectors, we can say the same thing like this: