30

I have been wondering about this question since I was in school. How can one number tell so much about the whole matrix being invertible or not?

I know the proof of this statement now. But I would like to know the intuition behind this result and why this result is actually true.


The proof I have in mind is:

If $A$ is invertible, then

$$ 1 = \det(I) = \det(AA^{-1}) = \det(A)\cdot\det(A^{-1})$$

whence $\det(A) \neq 0$.

Conversely, if $\det(A) \neq 0$, we have

$$ A adj(A) = adj(A)A = \det(A)I$$

whence $A$ is invertible.

$adj(A)$ is the adjugate matrix of $A$.

$$ adj(A)_{ji} = (-1)^{i+j}\det(A_{ij})$$

where $A_{ij}$ is the matrix obtained from $A$ by deleting $ith$ row and $jth$ column.

Any other insightful proofs are also welcome.

Vishal Gupta
  • 6,486
  • 4
  • 29
  • 51
  • 12
    The way I tend to remember it is that the determinant gives you the scale factor associated with the transformation represented by the matrix. And any figure scaled to "zero" looks the same… so there's not really enough information left to invert the transformation. – Aaron Golden Sep 28 '13 at 07:18

7 Answers7

50

Here's an explanation for three dimensional space ($3 \times 3$ matrices). That's the space I live in, so it's the one in which my intuition works best :-).

Suppose we have a $3 \times 3$ matrix $\mathbf{M}$. Let's think about the mapping $\mathbf{y} = f(\mathbf{x}) = \mathbf{M}\mathbf{x}$. The matrix $\mathbf{M}$ is invertible iff this mapping is invertible. In that case, given $\mathbf{y}$, we can compute the corresponding $\mathbf{x}$ as $\mathbf{x} = \mathbf{M}^{-1}\mathbf{y}$.

Let $\mathbf{u}$, $\mathbf{v}$, $\mathbf{w}$ be 3D vectors that form the columns of $\mathbf{M}$. We know that $\det{\mathbf{M}} = \mathbf{u} \cdot (\mathbf{v} \times \mathbf{w})$, which is the volume of the parallelipiped having $\mathbf{u}$, $\mathbf{v}$, $\mathbf{w}$ as its edges.

Now let's consider the effect of the mapping $f$ on the "basic cube" whose edges are the three axis vectors $\mathbf{i}$, $\mathbf{j}$, $\mathbf{k}$. You can check that $f(\mathbf{i}) = \mathbf{u}$, $f(\mathbf{j}) = \mathbf{v}$, and $f(\mathbf{k}) = \mathbf{w}$. So the mapping $f$ deforms (shears, scales) the basic cube, turning it into the parallelipiped with sides $\mathbf{u}$, $\mathbf{v}$, $\mathbf{w}$.

Since the determinant of $\mathbf{M}$ gives the volume of this parallelipiped, it measures the "volume scaling" effect of the mapping $f$. In particular, if $\det{\mathbf{M}} = 0$, this means that the mapping $f$ squashes the basic cube into something flat, with zero volume, like a planar shape, or maybe even a line. A "squash-to-flat" deformation like this can't possibly be invertible because it's not one-to-one --- several points of the cube will get "squashed" onto the same point of the deformed shape. So, the mapping $f$ (or the matrix $\mathbf{M}$) is invertible if and only if it has no squash-to-flat effect, which is the case if and only if the determinant is non-zero.

bubba
  • 40,254
  • 3
  • 55
  • 107
  • 5
    I signed up just so I could upvote this. Thank you. Very informative and intuitive. – Matt Klein Oct 09 '13 at 02:37
  • Thanks, Matt. I'm glad it helped. – bubba Oct 09 '13 at 06:14
  • nice explanation ! +1 only for "3D is where we live, so it's where our intuition works best" I was wondering what would it be if we live in higher dimensional space ? :) – Houda Jun 08 '15 at 20:27
15

I know this is pretty old, but for the people who might find this in a google search (I know I did), I thought I'd add this.

Remember that the space of all $n \times n$ matrices is isomorphic to the space of all operators on an $n$-dimensional vector space. In other words, the matrix is just some linear operator. Recall that the determinant is the product of the eigenvalues. Both $ \mathbb{C} $ and $ \mathbb{R} $ are integral domains so if the determinant is $0$ then that means you have a zero eigenvalue. That means, if your matrix is $A$, there exists a vector $0 \neq \overline{v} $ such that $ (A-0I)\overline{v}=\overline{0} $ which means $ A\overline{v}=\overline{0} $. Clearly $ \overline{0} $ gets sent to $ \overline{0} $ but so does some other non-zero vector. So the transformation isn't injective and, thus, non-invertible. This is just the intuition though.

Ludolila
  • 2,846
  • 6
  • 23
  • 31
Tom H.
  • 191
  • 1
  • 7
  • 5
    This is not just intuition. This is a rigorous proof and a very good one at that! – Vishal Gupta Jul 08 '15 at 23:57
  • @VishalGupta With the caveat that linear maps $\mathbb{R}^n\to \mathbb{R}^n$ might not have any eigenvalues at all. For instance the linear map on $\mathbb{R}^2$ or rotation by $\pi/2$. – Jamie Radcliffe Jun 13 '21 at 13:00
4

The absolute value of the determinant of a matrix is the volume of the parallelepiped spanned by the column vectors of that matrix.

Michael

Michael Hoppe
  • 16,684
  • 3
  • 28
  • 47
  • Then I would beg to ask for an intuition as to why the determinant is the volume and also how the invertibility of the matrix is related to the volume of the parellelopiped being non-zero. – Vishal Gupta Sep 28 '13 at 08:53
  • For the volume thing remember that the determinant of a diagonal matrix equals the product of its diagonal elements. For the invertibility the column vectors need to be lin. indep. What is the volume of a parallelepiped spanned by lin. dep. vectors? – Michael Hoppe Sep 28 '13 at 10:34
  • @Vishal [Relevant answer](http://math.stackexchange.com/questions/668/whats-an-intuitive-way-to-think-about-the-determinant/669#669) to determinant/volume. For the invertibility/volume think about the [nullspace](https://en.wikipedia.org/wiki/Kernel_(matrix)) of a non-invertible matrix. – jkn Sep 28 '13 at 10:39
2

Another classical way is more understandable: note that a determinant is not changed if we add one row to other and one column to another. Thus we obtain a diagonal matrix $B$. This matrix differs from $A$ by matrix-multipliers which correspond to elementary transformations and are invertible. So $A$ is invertible iff $B$ is invertible iff $\det(B) \neq 0$ iff $\det(A) \neq 0$.

Boris Novikov
  • 17,021
  • 1
  • 19
  • 31
1

I personally think of the determinant as a function $f(A)$ that has the following three properties:

  1. $f(AB) = f(A) f(B)$

  2. $f(T)$ = product of diagonals for triangular matrix $T$

  3. $f(E) \neq 0$ for an elementary matrix $E$

An elementary matrix $E$ is a matrix such that $EA$ either multiples a row, swaps a row, or swaps and adds a row of $A$.

Now Gaussian elimination is the process of applying elementary matrixes to $A$ and obtaining a upper triangular matrix. That is, every matrix can be written as $$ A = (E_1 \cdots E_n)^{-1} T$$

Which means $$f(A) = f(E_1)^{-1} \cdots f(E_n)^{-1} f(T)$$ and is nonzero if and only if $f(T)$ is nonzero. But a matrix is invertible if and only if the gaussian eliminated form $T$ has nonzero elements in the diagonal. That is, exactly, that $f(T) \neq 0$.


In my mind, any other definitions of the determinant, such as the practical way of computing it with adjunct matrices and such, are not the "real" or "primary" definition of the determinant. We want the determinant function to satisfy the above three properties. We also probably want the determinant to be a polynomial in the entries of our matrix. This then should give us the method of computing the determinant.

abnry
  • 14,161
  • 2
  • 33
  • 75
1

With all of the elegant ways of thinking about the determinant, I fear it is sometimes not emphasized enough that the easiest way to discover the determinant is to just solve the linear system \begin{align} a_{11} x_1 + a_{12} x_2 &= b_1 \\ a_{21} x_1 + a_{22} x_2 &= b_2 \end{align} by hand. When you do this, out pops the determinant! And we see immediately that if the determinant is $0$ then our solution doesn't work (because then we would be dividing by $0$). A similar computation can be done for $3 \times 3$ systems, etc.

This is probably how the determinant was first discovered. It is so easy that a middle school student can do it, and at the moment of discovery we see that a nonzero determinant guarantees that $Ax = b$ has a unique solution.

littleO
  • 48,104
  • 8
  • 84
  • 154
0

Consider $A$, an $n$ x $n$ matrix to be a mapping from $R^n$ to $R^n$. If the determinant is zero, the columns of $A$ are linearly dependent. This means that the nullspace of $A$ is nonempty. Hence the linear mapping $A$ is non-invertible, since several $x$ get mapped to the same value b, i.e there are multiple solutions (infinite in fact) to $Ax=b$ when a single solution exists. Example - both $x$ and $x+v$ get mapped to b, where $v$ is a vector in the nullspace of $A$

Sridhar Thiagarajan
  • 650
  • 2
  • 6
  • 21