For $n = 2$, I can visualize that the determinant $n \times n$ matrix is the area of the parallelograms by actually calculating the area by coordinates. But how can one easily realize that it is true for any dimensions?
-
43blue1brown did a ridiculously good video on the intuition behind why the determinant is the volume here – BlueRaja - Danny Pflughoeft Apr 20 '20 at 05:54
-
10@BlueRaja But he proved it only for $2\times 2$ matrices. And the $2\times 2$ case is just the case OP understands. They want to know the connection for $n\times n$ matrices. – Redbox Oct 06 '20 at 13:09
6 Answers
If the column vectors are linearly dependent, both the determinant and the volume are zero. So assume linear independence. The determinant remains unchanged when adding multiples of one column to another. This corresponds to a skew translation of the parallelepiped, which does not affect its volume. By a finite sequence of such operations, you can transform your matrix to diagonal form, where the relation between determinant (=product of diagonal entries) and volume of a "rectangle" (=product of side lengths) is apparent.

- 122,002

- 374,180
-
1thanks. from an constructive view of point, can the skew translation does not change volume plus multilinear condition fully determine the form of determinant? – ahala Jun 23 '13 at 15:34
-
2Yes. In abstract language: The vector space of alternating $n$-forms is one-dimensional – Hagen von Eitzen May 02 '14 at 11:58
-
I don't think the relation is apparent, even in two dimensions. The proof by picture in 2d takes some (pretty little) movement of equal areas around for the second shear. – Mitch Oct 19 '18 at 19:52
-
4This proof proves "The determinant is a reasonable definition for volume because it preserves its value under shear transformation for parallelipeds, just like volume". It does not show that the determinant equals volume. To show that, we need an analytical geometry expression for the volume of a paralleliped arrived from other principles (without using the determinant). Then we show that the determinant equals this definition of the volume. "The determinant equals the volume" is more a definition than something that can be proved. – user3180 Sep 09 '20 at 02:22
-
Another way to describe the problem: "the relation between determinant and volume is apparent". This is only true when the determinant is applied over orthogonal set of vectors. By no means can you generalize this to non-orthogonal vectors for free! – user3180 Sep 09 '20 at 02:58
-
Just because the function gives the correct answer for one input-output pair, does not mean it gives the right answer for all input-output pairs. – user3180 Sep 09 '20 at 02:59
-
If you argument is, volume is by definition any property that obeys invariance to skew transformations. Then the proof is complete. By we can trivially produce a counterexample, where a function is invariant to skew transformations and is not a volume. – user3180 Sep 09 '20 at 03:00
-
Your argument is that any function that 1) produces a value for an orthogonal set of input vectors that is equivalent to our analytical geometry expression for the volume of a cuboid, which is the product of the lengths and 2) produces the same value if the orthogonal input vectors are skewed must be the parallelipiped volume function, which works for arbitrary vectors. But we must rigorously prove that no other function exists that can satisfy these two conditions and not be the volume function. – user3180 Sep 09 '20 at 03:08
-
I think it is worth mentioning that columns switching is needed in general to reach a diagonal form. This would introduce a (-1)^k factor (with 0<=k<=n) to the value of the original determinant. This is in fact expected, and is related to the fact that the volume form requires the choice of an orientation. – Guillermo BCN Oct 14 '21 at 07:54
-
3@user3180 The argument is complete: if we accept that 1) For a cuboid the volume is given by the (absolute value) of the determinant of the corresponding diagonal matrix (multiplying the lengths of the edges), 2) shear operations don't change volume, 3) any matrix can be converted to the diagonal form by such operations; then the obvious consequence is that the determinant always gives the volume (not just for cuboids). I do however think it would be nice to justify point 2). How do we know shear operations leave volumes unchanged (even when for lower-dimension cases it is visually obvious)? – Guillermo BCN Oct 14 '21 at 08:28
Here is the same argument as Muphrid's, perhaps written in an elementary way.
Apply Gram-Schmidt orthogonalization to $\{v_{1},\ldots,v_{n}\}$, so that \begin{eqnarray*} v_{1} & = & v_{1}\\ v_{2} & = & c_{12}v_{1}+v_{2}^{\perp}\\ v_{3} & = & c_{13}v_{1}+c_{23}v_{2}+v_{3}^{\perp}\\ & \vdots \end{eqnarray*} where $v_{2}^{\perp}$ is orthogonal to $v_{1}$; and $v_{3}^{\perp}$ is orthogonal to $span\left\{ v_{1},v_{2}\right\} $, etc.
Since determinant is multilinear, anti-symmetric, then \begin{eqnarray*} \det\left(v_{1},v_{2},v_{3},\ldots,v_{n}\right) & = & \det\left(v_{1},c_{12}v_{1}+v_{2}^{\perp},c_{13}v_{1}+c_{23}v_{2}+v_{3}^{\perp},\ldots\right)\\ & = & \det\left(v_{1},v_{2}^{\perp},v_{3}^{\perp},\ldots,v_{n}^{\perp}\right)\\ & = & \mbox{signed volume}\left(v_{1},\ldots,v_{n}\right) \end{eqnarray*}

- 1,362
-
3Very nice explanation. Just an addition for the last equality. If $V$ is the matrix with columns $v_{1},v_{2},v_{3},\ldots,v_{n}$ then, because the columns are othogonal to each other, $U U^T = diag(||v_1||^2, \ldots, ||v_2||^2) $. Thus $|U U^T| = ||v_1||^2 \ldots ||v_n||^2$, which, given that $|U U^T| = |U| |U^T| = |U|^2$ yields $ |U| = ||v_1|| \ldots ||v_n|| = \mbox{signed volume}\left(v_{1},\ldots,v_{n}\right) $. – Sotiris Mar 09 '19 at 22:47
-
2@Sotiris I don't find this proof complete. If we combine Sotiris addendum with James proof, we end up showing that the $det(v_1, ..., v_n)$ = signed volume $(v_1, v_2^\perp, ... v_n^\perp)$ which is not complete – user3180 Aug 17 '20 at 17:30
-
1That is we don't show $det(v_1, ..., v_n)$ = signed volume $(v_1, ... v_n)$ – user3180 Aug 17 '20 at 17:39
-
1@user3180, https://math.stackexchange.com/questions/4154668/determinant-as-the-volume-of-a-box-in-n-dimensions – Kashmiri May 29 '21 at 05:39
In 2d, you calculate the area of a parallelogram spanned by two vectors using the cross product. In 3d, you calculate the volume of a parallelepiped using the triple scalar product. Both of these can be written in terms of a determinant, but it's probably not clear to you what the proper generalization is to higher dimensions.
That generalization is called the wedge product. Given $n$ vectors $v_1, v_2, \ldots, v_n$, the wedge product $v_1 \wedge v_2 \wedge \ldots \wedge v_n$ is called an $n$-vector, and it has as its magnitude the $n$-volume of that $n$-parallelepiped.
What is the relationship between the wedge product and the determinant? Quite simple, actually. There is a natural generalization of linear maps to work on $k$-vectors. Given a linear map $\underline T$ (which can be represented as a matrix), the action of that map on a $k$-vector is defined as
$$\underline T(v_1 \wedge v_2 \wedge \ldots \wedge v_k) \equiv \underline T(v_1) \wedge \underline T(v_2) \wedge \ldots \wedge \underline T(v_k)$$
When talking about $n$-vectors in an $n$-dimensional space, it's important to realize that the "vector space" of these $n$-vectors is in fact one-dimensional. That is, if you think about volume, there is only one such unit volume in a given space, and all other volumes are just scalar multiples of it. Hence, when we talk about the action of a linear map on an $n$-vector, we can see that
$$\underline T(v_1 \wedge v_2 \wedge \ldots \wedge v_n) = \alpha [v_1 \wedge v_2 \wedge \ldots \wedge v_n]$$
for some scalar $\alpha$. In fact, this is a coordinate system independent definition of the determinant!
When you build a matrix out of $n$ vectors $f_1, f_2, \ldots, f_n$ as the matrix's columns, what you're really doing is the following: you're saying that, if you have a basis $e_1, e_2, \ldots, e_n$, then you're defining a map $\underline T$ such that $\underline T(e_1) = f_1$, $\underline T(e_2) = f_2$, and so on. So when you input $e_1 \wedge e_2 \wedge \ldots \wedge e_n$, you get
$$\underline T(e_1 \wedge e_2 \wedge \ldots \wedge e_n) = (\det \underline T) e_1 \wedge e_2 \wedge \ldots \wedge e_n= f_1 \wedge f_2 \wedge \ldots \wedge f_n$$
This is how you can use a matrix determinant to calculate volumes: it's just an easy way of constructing something that automatically computes the wedge product.
Edit: how one can see that the wedge product accurately gives the volume of a parallelepiped. Any vector can be broken down into perpendicular and parallel parts with respect to another vector, to a plane, and so on (or to any $k$-vector). As such, if I have two vectors $a$ and $b$, then the wedge product $a \wedge b = a \wedge b_\perp$, where $b_\perp$ is effectively the height of the parallelogram. Similarly, if I construct a parallelepiped with a vector $c$, then the wedge product $a \wedge b \wedge c = (a \wedge b_\perp) \wedge c_\perp$, where $c_\perp$ lies entirely normal to $a \wedge b_\perp$. So we can recursively do this for any $k$-vector, looking at orthogonal vectors instead, which is much simpler to see the volumes from.

- 19,902
-
Thanks. Actually my question raised exactly from learning about wedge product. How can one see that the wedge product has its magnitude the n-volume of that n-parallelepiped? I guess it is equivalent to ask in term of determinant as in my question. – ahala Jun 23 '13 at 15:39
-
-
@Muphrid Can you please friendly remind me why $T(e_1 \wedge \cdots \wedge e_n) = det(T) e_1 \wedge \cdots \wedge e_n$? Also, $a \wedge b = a \wedge b_\perp$. – James C Mar 16 '21 at 08:03
The determinant of a matrix A is the unique function that satisfies:
- $\det(A)=0$ when two columns are equal
- the determinant is linear in the columns
- if A is the identity $\det(A)=1$.
You can easily convince yourself that the oriented volume $\operatorname{vol}(v_1,v_2,\ldots,v_n)$ between $v_1, v_2,\ldots, v_n$ vectors is a function that satisfies exactly the same properties if we place the vectors as the columns of a matrix $A=(v_1,\ldots,v_n)$. Hence $\operatorname{vol}(v_1,v_2,\ldots,v_n)=\det(A)$.

- 427,504

- 185
-
2one has to be convinced why such function is unique to reach the conclusion. – ahala Dec 12 '16 at 14:37
-
1True. You can prove that the determinant is unique by constructing it using Gaussian elimination as the signed product of all the pivot of a matrix. – jacopoviti Dec 13 '16 at 14:59
Determinant involves a cross-product of the first two vectors and a dot of the result with the third. The result of a cross product is a vector whose magnitude is the area of its null space. Said simply, any plane in 3D is the null space of its normal.The size of the plane is defined by the length of the normal. The volume is found by projecting this normal onto the third vector.

- 21
-
-
Yes, my bad. Assuming from nD we get an nxn matrix. Remove one row and then find the null space of the remaining n-1xn hyperplane which is inevitably a vector perpendicular to that hyperplane. Dot this vector with the one removed and I believe this amounts to the volume in nD. – Shadi Dec 26 '16 at 03:14
You can also invoke the change-of-variable theorem in higher dimensions. An $n$-dimensional parallelepiped $\mathcal P=\mathcal P(a_1,\dots,a_n)$ in $\mathbb R^n$ (where the $a_i$ are independent vectors in $\mathbb R^n$) is the set of all $x$ such that: $$ x=c_1a_1+\dots+c_ka_n, $$ with $0\leq c_i\leq 1$. We can define the linear transformation $h(x)=A\cdot x$, where $A$ is the $n\times n$ matrix with $a_i$ as its columns. This gives us $\mathcal P=h([0,1]^n)$. The volume of $h([0,1]^n)$ is equal to $h((0,1)^n))$ (those sets are equal modulo a set of measure zero), so we can apply the change-of-variable theorem: $$ v(\mathcal P)=\int_{h((0,1)^n)}1=\int_{(0,1)^n}\vert\det Dh\vert=\vert\det A\vert. $$
Edit:
As pointed out in the comments, conceptually this argument is circular. Alternatively, one can give a proof that the $n$-dimensional Lebesgue measure $\lambda^n$ transforms in this way, i.e.:
Given an (invertible) linear map $A:\mathbb R^n\to\mathbb R^n$, we have $$ \lambda^n(A(B))=\vert\det A\vert\lambda^n(B) $$ for any Borel measurable set $B$ of $\mathbb R^n$.
The proof uses the push-forward measure $\lambda^n(A(-))$ and invokes the uniqueness${}^{(1)}$ of the Lebesgue measure, along with some elementary linear algebra: first show that the transformation rule holds for diagonal matrices and orthogonal matrices, and then use that any invertible matrix $A$ can be written as $UDV$ where $U,V$ are orthogonal and $D$ is diagonal.
${}^{(1)}$ For any measure $\mu$ on $\mathbb R^n$ such that $\mu([0,1]^n)<\infty$ we have $\mu=\mu([0,1]^n)\cdot\lambda^n$.
For the details see Schilling's book Measures, Integrals and Martingales (Appendix C The Volume of a Parallelepiped).

- 3,960
- 4
- 19
- 37
-
7Isn't the change of variables theorem based on what we are asked to prove? – Theorem Aug 16 '18 at 14:56