Section D Dimension

Almost every vector space we have encountered has been infinite in size (an exception is Example VSS). But some are bigger and richer than others. Dimension, once suitably defined, will be a measure of the size of a vector space, and a useful tool for studying its properties. You probably already have a rough notion of what a mathematical definition of dimension might be — try to forget these imprecise ideas and go with the new ones given here.

Subsection D Dimension

Definition D Dimension

Suppose that $V$ is a vector space and $\set{\vectorlist{v}{t}}$ is a basis of $V$. Then the dimension of $V$ is defined by $\dimension{V}=t$. If $V$ has no finite bases, we say $V$ has infinite dimension.

This is a very simple definition, which belies its power. Grab a basis, any basis, and count up the number of vectors it contains. That is the dimension. However, this simplicity causes a problem. Given a vector space, you and I could each construct different bases — remember that a vector space might have many bases. And what if your basis and my basis had different sizes? Applying Definition D we would arrive at different numbers! With our current knowledge about vector spaces, we would have to say that dimension is not “well-defined.” Fortunately, there is a theorem that will correct this problem.

In a strictly logical progression, the next two theorems would precede the definition of dimension. Many subsequent theorems will trace their lineage back to the following fundamental result.

Theorem SSLD Spanning Sets and Linear Dependence

Suppose that $S=\set{\vectorlist{v}{t}}$ is a finite set of vectors which spans the vector space $V$. Then any set of $t+1$ or more vectors from $V$ is linearly dependent.

Notice how the swap of the two summations is so much easier in the third step above, as opposed to all the rearranging and regrouping that takes place in the previous proof. And using only about half the space. And there are no ellipses (…).

Theorem SSLD can be viewed as a generalization of Theorem MVSLD. We know that $\complex{m}$ has a basis with $m$ vectors in it (Theorem SUVB), so it is a set of $m$ vectors that spans $\complex{m}$. By Theorem SSLD, any set of more than $m$ vectors from $\complex{m}$ will be linearly dependent. But this is exactly the conclusion we have in Theorem MVSLD. Maybe this is not a total shock, as the proofs of both theorems rely heavily on Theorem HMVEI. The beauty of Theorem SSLD is that it applies in any vector space. We illustrate the generality of this theorem, and hint at its power, in the next example.

Example LDP4 Linearly dependent set in $P_4$

Theorem SSLD is indeed powerful, but our main purpose in proving it right now was to make sure that our definition of dimension (Definition D) is well-defined. Here is the theorem.

Theorem BIS Bases have Identical Sizes

Suppose that $V$ is a vector space with a finite basis $B$ and a second basis $C$. Then $B$ and $C$ have the same size.

Theorem BIS tells us that if we find one finite basis in a vector space, then they all have the same size. This (finally) makes Definition D unambiguous.

Subsection DVS Dimension of Vector Spaces

We can now collect the dimension of some common, and not so common, vector spaces.

Theorem DCM Dimension of $\complex{m}$

The dimension of $\complex{m}$ (Example VSCV) is $m$.

Theorem DP Dimension of $P_n$

The dimension of $P_{n}$ (Example VSP) is $n+1$.

Theorem DM Dimension of $M_{mn}$

The dimension of $M_{mn}$ (Example VSM) is $mn$.

Example DSM22 Dimension of a subspace of $M_{22}$
Example DSP4 Dimension of a subspace of $P_4$
Example DC Dimension of the crazy vector space

It is possible for a vector space to have no finite bases, in which case we say it has infinite dimension. Many of the best examples of this are vector spaces of functions, which lead to constructions like Hilbert spaces. We will focus exclusively on finite-dimensional vector spaces. OK, one infinite-dimensional example, and then we will focus exclusively on finite-dimensional vector spaces.

Example VSPUD Vector space of polynomials with unbounded degree
Sage D Dimension

Subsection RNM Rank and Nullity of a Matrix

For any matrix, we have seen that we can associate several subspaces — the null space (Theorem NSMS), the column space (Theorem CSMS), row space (Theorem RSMS) and the left null space (Theorem LNSMS). As vector spaces, each of these has a dimension, and for the null space and column space, they are important enough to warrant names.

Definition NOM Nullity Of a Matrix

Suppose that $A$ is an $m\times n$ matrix. Then the nullity of $A$ is the dimension of the null space of $A$, $\nullity{A}=\dimension{\nsp{A}}$.

Definition ROM Rank Of a Matrix

Suppose that $A$ is an $m\times n$ matrix. Then the rank of $A$ is the dimension of the column space of $A$, $\rank{A}=\dimension{\csp{A}}$.

Example RNM Rank and nullity of a matrix

There were no accidents or coincidences in the previous example — with the row-reduced version of a matrix in hand, the rank and nullity are easy to compute.

Theorem CRN Computing Rank and Nullity

Suppose that $A$ is an $m\times n$ matrix and $B$ is a row-equivalent matrix in reduced row-echelon form. Let $r$ denote the number of pivot columns (or the number of nonzero rows). Then $\rank{A}=r$ and $\nullity{A}=n-r$.

Every archetype (Archetypes) that involves a matrix lists its rank and nullity. You may have noticed as you studied the archetypes that the larger the column space is the smaller the null space is. A simple corollary states this trade-off succinctly. (See Proof Technique LC.)

Theorem RPNC Rank Plus Nullity is Columns

Suppose that $A$ is an $m\times n$ matrix. Then $\rank{A}+\nullity{A}=n$.

When we first introduced $r$ as our standard notation for the number of nonzero rows in a matrix in reduced row-echelon form you might have thought $r$ stood for “rows.” Not really — it stands for “rank”!

Sage RNM Rank and Nullity of a Matrix

Subsection RNNM Rank and Nullity of a Nonsingular Matrix

Let us take a look at the rank and nullity of a square matrix.

Example RNSM Rank and nullity of a square matrix

The value of either the nullity or the rank are enough to characterize a nonsingular matrix.

Theorem RNNM Rank and Nullity of a Nonsingular Matrix

Suppose that $A$ is a square matrix of size $n$. The following are equivalent.

  1. A is nonsingular.
  2. The rank of $A$ is $n$, $\rank{A}=n$.
  3. The nullity of $A$ is zero, $\nullity{A}=0$.

With a new equivalence for a nonsingular matrix, we can update our list of equivalences (Theorem NME5) which now becomes a list requiring double digits to number.

Theorem NME6 Nonsingular Matrix Equivalences, Round 6

Suppose that $A$ is a square matrix of size $n$. The following are equivalent.

  1. $A$ is nonsingular.
  2. $A$ row-reduces to the identity matrix.
  3. The null space of $A$ contains only the zero vector, $\nsp{A}=\set{\zerovector}$.
  4. The linear system $\linearsystem{A}{\vect{b}}$ has a unique solution for every possible choice of $\vect{b}$.
  5. The columns of $A$ are a linearly independent set.
  6. $A$ is invertible.
  7. The column space of $A$ is $\complex{n}$, $\csp{A}=\complex{n}$.
  8. The columns of $A$ are a basis for $\complex{n}$.
  9. The rank of $A$ is $n$, $\rank{A}=n$.
  10. The nullity of $A$ is zero, $\nullity{A}=0$.

Sage NME6 Nonsingular Matrix Equivalences, Round 6