Académique Documents
Professionnel Documents
Culture Documents
Introduction
In this short chapter we examine more generally the concept of orthogonality, which
has already been encountered in our work on orthogonal diagonalisation.
71
It is important to realise that the inner product is just a number, not another vector
or a matrix.
It is a simple matter to check that the inner product defined above for real vectors
is indeed an inner product according to this more abstract definition, and we shall
call it the standard inner product on Rn . The abstract definition, though, applies
to more than just the vector space Rn , and there is some advantage in developing
results in terms of the general notion of inner product. If a vector space has an inner
product defined on it, we refer to it as an inner product space.
Example: Suppose that V is the vector space consisting of all real polynomial functions of degree at most n; that is, V consists of all functions of the form
p(x) = a0 + a1 x + a2 x2 + + an xn , where a0 , a1 , . . . , an R.
The addition and scalar multiplication are, as usual, defined pointwise. Let x1 , x2 , . . . , xn+1
be n fixed, different, real numbers, and define, for p, q V ,
hp, qi =
n+1
X
p(xi )q(xi ).
i=1
Then this is an inner product. To see this, we check the properties in the definition
of an inner product. Property (ii) is clear. For (i), we have
hp, pi =
n+1
X
p(xi )2 0.
i=1
Clearly, if p is the zero vector of the vector space (which is the identically-0 function),
then hp, pi = 0. To finish verifying (i) we need to check that if hp, pi = 0 then p must
be the zero function. Now, hp, pi = 0 must mean that p(xi ) = 0 for i = 1, 2, . . . , n + 1.
So p has n + 1 different roots. But p has degree no more than n, so p must be the
identically-zero function. (A non-zero polynomial of degree at most n has no more
than n distinct roots.) Part (iii) is left to you:
72
Orthogonality
Orthogonal vectors
We have already said (in the discussion of orthogonal diagonalisation) what it means
for two vectors x, y in Rn to be orthogonal: it means that xT y = 0. In other words,
x, y are orthogonal if hx, yi = 0. We take this as the general definition of orthogonality
in an inner product space:
Definition 5.1 Suppose that V is an inner product space. Then x, y V are said to
be orthogonal if hx, yi = 0. We write x y to mean that x, y are orthogonal.
Example: With the usual inner product on R3 , the vectors x = (1, 1, 0)T and
y = (2, 2, 3)T are orthogonal.
Activity 5.3 Check this!
Geometrical interpretation
A geometrical interpretation can be given to the notion of orthogonality in Rn . Consider a very simple example with n = 2. Suppose that x = (1, 1)T and y = (1, 1)T .
Then x, y are orthogonal, as is easily seen. We can represent x, y geometrically on the
standard two-dimensional (x, y)-plane: x is represented as an arrow from the origin
(0, 0) to the point (1, 1); and y is represented as an arrow from the origin to the point
(1, 1). This is shown in the figure. It is clear that these arrowsthe geometrical
interpretations of x, yare at right angles to each other: they are perpendicular.
73
y
6
(1, 1)
(1, 1)
@
I
@
@
6
@
@
@
@
@
@
(0, 0)
Theorem 5.1 Suppose that V is an inner product space and that vectors v1 , v2 , . . . , vk
V are pairwise orthogonal (vi vj for i 6= j), and none is the zero-vector. Then
{v1 , v2 , . . . , vk } is a linearly independent set of vectors.
Proof We need to show that if
1 v1 + 2 v2 + + k vk = 0,
(the zero-vector), then 1 = 2 = = k = 0. Let i be any integer between 1 and
k. Then
hvi , 1 v1 + 2 v2 + + k vk i = hvi , 0i = 0.
But, since hvi , vj i = 0 for j 6= i,
hvi , 1 v1 +2 v2 + +k vk i = 1 hvi , v1 i+2 hvi , v2 i+ +k hvi , vk i = i hvi , vi i = i kvi k2 .
So we have i kvi k2 = 0. Since vi 6= 0, kvi k2 6= 0 and hence i = 0. But i was any
integer in the range 1 to n, so we deduce that
1 = 2 = . . . = k = 0,
as required.
74
T
x1 x1 xT1 x2 xT1 x3
1 0 0
x1
0 1 0 = I = P T P = xT2 ( x1 x2 x3 ) = xT2 x1 xT2 x2 xT2 x3 .
xT3 x1 xT3 x2 xT3 x3
0 0 1
xT3
But, if i 6= j, xTi xj = 0 means precisely that the columns xi , xj are orthogonal. The
second statement is that kxi k2 = 1, which means (since kxi k 0) that kxi k = 1;
that is, xi is of length 1. This indicates the following characterisation: a matrix P
is orthogonal if and only if, as vectors, its columns are pairwise orthogonal, and each
has length 1.
When a set of vectors {x1 , x2 , . . . , xk } is such that any two are orthogonal and,
furthermore, each has length 1, we say that the vectors form an orthonormal set
(ONS) of vectors. So we can restate our previous observation as follows.
Proof Let x, y be any two vectors of V . For any real number , we consider the
vector x + y. Certainly, kx + yk2 0 for all . But
kx + yk2
= hx + y, x + yi
= 2 hx, xi + hx, yi + hy, xi + hy, yi
= 2 kxk2 + 2hx, yi + kyk2 .
75
(Recall that |hx, yi| denotes the absolute value of the inner product.)
For example, if we take V to be Rn and consider the standard inner product on Rn ,
then for all x, y Rn , the Cauchy-Schwarz inequality tells us that
v
v
u n
n
n
X
u
X
uX
u
2
t
xi t
yi2 .
xi yi
i=1
i=1
i=1
=
=
=
=
=
hx + y, x + yi
hx, x + yi + hy, x + yi
hx, xi + hx, yi + hy, xi + hy, yi
kxk2 + 2hx, yi + kyk2
kxk2 + kyk2 ,
where the last line follows from the fact that, x, y being orthogonal, hx, yi = 0.
= hx + y, x + yi
= hx, x + yi + hy, x + yi
76
v1
.
kv1 k
Then we define
u2 = v2 hv2 , e1 ie1 ,
and set
e2 =
u2
.
ku2 k
Next, we define
u3 = v3 hv3 , e1 ie1 hv3 , e2 ie2
and set
e3 =
u3
.
ku3 k
i
X
j=1
ui+1
.
kui+1 k
It turns out that the resulting set {e1 , e2 , . . . , ek } has the required properties.
Example: In R4 , let us find an orthonormal basis for the linear span of the three
vectors
v1 = (1, 1, 1, 1)T , v2 = (1, 4, 4, 1)T , v3 = (4, 2, 2, 0).
77
First, we have
e1 =
v1
v1
1
=
= v1 = (1/2, 1/2, 1/2, 1/2)T .
2
2
2
2
kv1 k
2
1 +1 +1 +1
Next, we have
u2 = v2 hv2 , e1 ie1 = (1, 4, 4, 1)T (3)(1/2, 1/2, 1/2, 1/2)T = (5/2, 5/2, 5/2, 5/2)T ,
and we set
e2 =
u2
= (1/2, 1/2, 1/2, 1/2).
ku2 k
(Note: to do this last step, I merely noted that a normalised vector in the same
direction as u2 is also a normalised vector in the same direction as (1, 1, 1, 1)T ,
and this second vector is easier to work with.) Continuing, we have
u3
Then,
e3 =
So
u3
= (1/2, 1/2, 1/2, 1/2)T .
ku3 k
1/2
1/2
1/2
1/2
1/2
1/2
1/2
Activity 5.4 Verify that the set {e1 , e2 , e3 } of this example is an orthonormal set.
Learning outcomes
At the end of this chapter and the relevant reading, you should be able to:
explain what is meant by an inner product on a vector space
verify that a given inner product is indeed an inner product
compute norms in inner product spaces
explain why orthogonality of a set of vectors implies linear independence
explain what is meant by an orthonormal set of vectors
explain why an nn matrix is orthogonal if and only if it possesses an orthonormal set of n eigenvectors
know and apply the Cauchy-Schwarz inequality, the Generalised Pythagoras
Theorem, and the triangle inequality for norms
use the Gram-Schmidt orthonormalisation process
n
m X
X
aij bij .
i=1 j=1
Question 5.4 Use the Gram-Schmidt process to find an orthonormal basis for the
subspace of R4 spanned by the vectors
1
1
0
0
2
1
v1 = , v2 = , v3 = .
1
1
2
0
1
1
m X
n
X
a2ij 0,
i=1 j=1
and this equals zero if and only if for every i and every j, aij = 0, which means that
A is the zero matrix, which in this vector space is the zero vector. Property (ii) is
easy to verify, as also is (iii).
= hx + y, x + yi + hx y, x yi
= hx, xi + 2hx, yi + hy, yi + hx, xi 2hx, yi + hy, yi
= 2hx, xi + 2hy, yi
= 2kxk2 + 2kyk2 .
0
u2
1 2
e2 =
= .
ku2 k
5 0
1
80
Next,
1
1/5
u3 = v3 hv3 , e2 ie2 hv3 , e1 ie1 = =
.
1
2/5
Normalising u3 we obtain
1
e3 = (5, 1, 5, 2)T .
55
The required basis is {e1 , e2 , e3 } .
81