Folioby Interconnected
Log InSign Up

Inner Product Spaces and Orthonormal Bases

The Cauchy–Schwarz inequality |<u,v>| <= ||u|| ||v|| is the cornerstone of inner product spaces. We prove it from the axioms, then develop the Gram–Schmidt process for constructing orthonormal bases, orthogonal complements, and orthogonal projections that give the closest point in a subspace.

FO
Folio Official
March 1, 2026

1. The Definition of an Inner Product

Definition 1 (Inner product).
A map ⟨⋅,⋅⟩:V×V→R on a real vector space V is called an inner product if it satisfies the following three axioms:
  1. Symmetry.⟨u,v⟩=⟨v,u⟩ for all u,v∈V.

  2. Linearity in the first argument.⟨au+bv,w⟩=a⟨u,w⟩+b⟨v,w⟩ for all a,b∈R and u,v,w∈V.

  3. Positive definiteness.⟨v,v⟩≥0 for all v∈V, with equality if and only if v=0.

A vector space equipped with an inner product is called an inner product space.
Definition 2 (Norm and orthogonality).
In an inner product space V, the norm (or length) of a vector v is defined by ∥v∥=⟨v,v⟩​. Two vectors u,v are said to be orthogonal if ⟨u,v⟩=0; we write u⊥v.
Example 3.
  • The standard inner product on Rn: ⟨x,y⟩=∑i=1n​xi​yi​=xTy.

  • An inner product on C[a,b]: ⟨f,g⟩=∫ab​f(x)g(x)dx.

  • An inner product on Mn​(R): ⟨A,B⟩=tr(ATB).

2. The Cauchy–Schwarz Inequality

Theorem 4 (Cauchy–Schwarz inequality).
For any vectors u,v in an inner product space V,
∣⟨u,v⟩∣≤∥u∥⋅∥v∥.
Equality holds if and only if u and v are linearly dependent.
Proof.
If v=0, both sides vanish and the inequality is trivially satisfied. Assume v=0. For any t∈R, positive definiteness gives
0≤∥u−tv∥2=⟨u−tv,u−tv⟩=∥u∥2−2t⟨u,v⟩+t2∥v∥2.
Substituting t=⟨u,v⟩/∥v∥2 yields
0≤∥u∥2−∥v∥2⟨u,v⟩2​,
and rearranging gives ⟨u,v⟩2≤∥u∥2∥v∥2. □
Theorem 5 (Triangle inequality).
For any vectors u,v in an inner product space, ∥u+v∥≤∥u∥+∥v∥.
Proof.
We compute
∥u+v∥2=∥u∥2+2⟨u,v⟩+∥v∥2≤∥u∥2+2∥u∥∥v∥+∥v∥2=(∥u∥+∥v∥)2,
where the inequality follows from ⟨u,v⟩≤∣⟨u,v⟩∣≤∥u∥∥v∥ (the Cauchy–Schwarz inequality). Since both sides are nonnegative, taking square roots gives the result. □

3. Orthonormal Bases

Definition 6 (Orthonormal system and orthonormal basis).
A set {e1​,…,en​} is called an orthonormal system if
⟨ei​,ej​⟩=δij​={10​if i=j,if i=j.​
An orthonormal system that is also a basis is called an orthonormal basis (ONB).
Theorem 7.
Every orthonormal system is linearly independent.
Proof.
Suppose ∑i​ci​ei​=0. Taking the inner product with ek​ gives ck​=⟨∑i​ci​ei​,ek​⟩=⟨0,ek​⟩=0. □
Theorem 8 (Fourier expansion).
If {e1​,…,en​} is an orthonormal basis for V, then every v∈V admits the expansion
v=i=1∑n​⟨v,ei​⟩ei​.
In other words, the i-th coordinate of v with respect to the orthonormal basis is simply ⟨v,ei​⟩.
Proof.
Since {e1​,…,en​} is a basis, we may write v=∑i=1n​ai​ei​ uniquely. Taking the inner product of both sides with ek​ gives ⟨v,ek​⟩=∑i=1n​ai​⟨ei​,ek​⟩=∑i=1n​ai​δik​=ak​. □

4. The Gram–Schmidt Process

Theorem 9 (Gram–Schmidt orthonormalization).
Let {v1​,…,vn​} be a linearly independent set in an inner product space V. The following procedure produces an orthonormal set {e1​,…,en​}:

  1. Set u1​=v1​ and e1​=u1​/∥u1​∥.

  2. For k=2,…,n: set uk​=vk​−∑j=1k−1​⟨vk​,ej​⟩ej​ and ek​=uk​/∥uk​∥.

Moreover, span{e1​,…,ek​}=span{v1​,…,vk​} at each stage.
Proof.
We proceed by induction on k. The case k=1 is clear. Suppose {e1​,…,ek−1​} is an orthonormal basis for span{v1​,…,vk−1​}. Define uk​=vk​−∑j=1k−1​⟨vk​,ej​⟩ej​. For each j≤k−1, we have ⟨uk​,ej​⟩=⟨vk​,ej​⟩−⟨vk​,ej​⟩=0. Since vk​ is not in span{v1​,…,vk−1​} (by linear independence), we have uk​=0. Setting ek​=uk​/∥uk​∥ completes the inductive step, and clearly span{e1​,…,ek​}=span{v1​,…,vk​}. □

5. Orthogonal Complements

Definition 10 (Orthogonal complement).
Let W be a subspace of an inner product space V. The orthogonal complement of W is
W⊥={v∈V∣⟨v,w⟩=0 for all w∈W}.
Theorem 11.
Let V be a finite-dimensional inner product space and W a subspace of V. Then:
  1. W⊥ is a subspace of V.

  2. V=W⊕W⊥.

  3. dimW⊥=dimV−dimW.

  4. (W⊥)⊥=W.

Proof.
(1) For u1​,u2​∈W⊥ and scalars a,b, we have ⟨au1​+bu2​,w⟩=a⟨u1​,w⟩+b⟨u2​,w⟩=0 for all w∈W, so au1​+bu2​∈W⊥.

(2) Apply the Gram–Schmidt process to obtain an orthonormal basis {e1​,…,ek​} for W. Given any v∈V, set w=∑i=1k​⟨v,ei​⟩ei​∈W. Then for each j≤k, ⟨v−w,ej​⟩=⟨v,ej​⟩−⟨v,ej​⟩=0, so v−w∈W⊥. Thus v=w+(v−w) gives V=W+W⊥. The intersection W∩W⊥={0} is immediate: if u∈W∩W⊥, then ⟨u,u⟩=0, hence u=0.

(3) From the direct sum decomposition, dimV=dimW+dimW⊥.

(4) The inclusion W⊆(W⊥)⊥ follows from the definition. By (3), dim(W⊥)⊥=dimV−dimW⊥=dimW, so W=(W⊥)⊥. □

6. Orthogonal Projection

Definition 12 (Orthogonal projection).
Given the decomposition V=W⊕W⊥, every v∈V can be written uniquely as v=w+w⊥ with w∈W and w⊥∈W⊥. The vector w is called the orthogonal projection of v onto W, denoted projW​(v)=w.
Theorem 13.
If {e1​,…,ek​} is an orthonormal basis for W, then
projW​(v)=i=1∑k​⟨v,ei​⟩ei​.
Proof.
Set w=∑i=1k​⟨v,ei​⟩ei​∈W. For each j=1,…,k,
⟨v−w,ej​⟩=⟨v,ej​⟩−i=1∑k​⟨v,ei​⟩⟨ei​,ej​⟩=⟨v,ej​⟩−⟨v,ej​⟩=0,
so v−w∈W⊥. The decomposition v=w+(v−w) is the direct-sum decomposition V=W⊕W⊥, whence projW​(v)=w. □
Theorem 14 (Best approximation theorem).
The projection projW​(v) is the closest point in W to v:
∥v−projW​(v)∥≤∥v−w∥for all w∈W,
with equality if and only if w=projW​(v).
Proof.
Let v^=projW​(v). For any w∈W,
∥v−w∥2=∥v−v^+v^−w∥2=∥v−v^∥2+∥v^−w∥2,
where the cross term vanishes because v−v^∈W⊥ and v^−w∈W are orthogonal. Since ∥v^−w∥2≥0, we obtain ∥v−w∥2≥∥v−v^∥2, with equality only when v^=w. □
Linear AlgebraAlgebraTextbookInner Product SpaceGram-SchmidtOrthogonal Complement
FO
Folio Official

Mathematics "between the lines" — exploring the intuition textbooks leave out, written in LaTeX on Folio.

1 followers·107 articles
Linear Algebra TextbookPart 10 of 13
Previous
Diagonalization: Simplifying Matrices by Choice of Basis
Next
Jordan Normal Form: Beyond Diagonalization

Share your expertise with the world

Write articles with LaTeX support, build your audience, and earn from your knowledge.

Start Writing — It's Free

More from Folio Official

Folio Official·March 1, 2026

Matrices and Representation of Linear Maps

Every linear map between finite-dimensional spaces is uniquely represented by a matrix once bases are chosen, and composition of maps corresponds to matrix multiplication. We derive the change-of-basis formula, characterize invertible matrices, and show that matrix rank equals the rank of the corresponding linear map.

Linear AlgebraAlgebraTextbook
Folio Official·March 1, 2026

The Determinant: A Scalar Invariant of Square Matrices

The determinant of a square matrix is the unique scalar-valued function characterized by alternation and multilinearity. We construct it via the Leibniz formula, prove the product formula det(AB) = det(A)det(B), and derive Cramer's rule for solving linear systems.

Linear AlgebraAlgebraTextbook
Folio Official·March 1, 2026

Jordan Normal Form: Beyond Diagonalization

Every complex square matrix is similar to a Jordan normal form — a block-diagonal matrix of Jordan blocks that is unique up to block ordering. We prove existence and uniqueness, connect the Jordan form to the minimal polynomial, and apply it to compute the matrix exponential e^{tA}.

Linear AlgebraAlgebraTextbook
Folio Official·March 24, 2026

The Singular Value Decomposition: Structure of Arbitrary Matrices

Every real m x n matrix factors as A = U Sigma V^T, where U and V are orthogonal and Sigma is diagonal — this is the singular value decomposition. We prove existence, show how the SVD yields optimal low-rank approximations (Eckart–Young theorem), and construct the Moore–Penrose pseudoinverse for least-squares solutions.

Linear AlgebraAlgebraTextbook