In this section, I'll derive another formula for determinants. Using this formula, we'll be able to show that there is only one function on matrices which satisfies the axioms for a determinant.
The main theorem says that a function on matrices which satisfies the first two axioms for a determinant function --- linearity on the rows, and alternating (0 on matrices with two equal rows) --- is a sum of products of entries of the matrix, "scaled" by the value of the function on the identity matrix ( ). Each product of entries is formed by choosing n entries, one from each row and each column, in all possible ways; the products get a plus or minus sign according to the sign of the permutation that the chosen entries "represent".
Theorem. Let R be a commutative ring with identity. If is linear in each row and is 0 on matrices with equal rows, then
Proof. This proof is a little complicated; if you wish to skip it for now, at least try to understand the statement of the theorem (and see the discussion that follows).
Write the matrix in terms of its rows:
Notice that the first row of A can be written this way:
Here is the j-th standard basis vector (equivalently, the j-th row of the identity matrix). So in summation notation,
I can use linearity applied to row 1 to expand the determinant of A into a sum of determinants:
In the same way,
Use linearity applied to row 2 to expand the determinant terms in the last D-sum:
Continue in this way for all n rows. I'll switch notation and use as the summation indices:
If two of the j's are equal, then the e's are equal --- e.g. if , then . But D is 0 on matrices with equal rows. So terms with two of the j's equal are 0. Hence, I only need to consider terms where all the j's are distinct numbers in the set . This means that is a permutation of . So I can just sum over all permutations in :
Consider the last matrix:
Each of the e's is a standard basis vector. But the standard basis vectors are the rows of the identity matrix. Thus, this matrix is just the identity matrix with its rows permuted by . Every permutation is a product of transpositions. By an earlier lemma, since D is linear on the rows and is 0 for matrices with equal rows, a transposition (i.e. a row swap) multiplies the value of D by -1. Hence,
Therefore,
Corollary. Let R be a commutative ring with identity, and let .
Proof. Since the determinant function defined by expansion by cofactors is alternating and linear in each row, it satisfies the conditions of the theorem. Since , the formula in the theorem becomes
I'll call this the permutation formula for the determinant.
Let's try to understand a typical product in the sum: . You can see that the row indices go from 1 to n, so we're choosing one entry of the matrix from each row. The column indices are , , ... . This is a permutation of , which means each of the numbers from 1 to n is chosen exactly once. This means that we're also choosing the entries so that one comes from each column. We're summing over all permutations of , so we're choosing entries for our products in all such ways. Let's see how this looks for small matrices.
Consider a matrix:
I have to choose 2 entries at a time, so the 2 entries come from different rows and columns. I multiply the 2 chosen entries to get one of the product terms. There are ways to do this; they are
Next, consider a matrix:
I have to choose 3 entries at a time, so the 3 entries come from different rows and columns. I multiply the 3 entries to get one of the product terms, and I do this in all possible ways. There are ways to do this; they are
For a matrix, there will be products!
The products we're getting will each get a plus or minus sign, depending on the permutation the product represents. We'll see how to determine the signs in the example below. Finally, all the signed products are added up to get the determinant.
Example. Use the permutation formula to compute the following real determinant:
I have to choose 3 entries from the matrix at a time, in such a way that there is one entry from each row and each column. For each such choice, I take the product of the three elements and multiply by the sign of the permutation of the elements, which I'll describe below. Finally, I add up the results.
In order to do this systematically, focus on the first column. I can choose 2, 1, or 5 from column 1.
If I choose 2 from column 1, I can either choose -1 from column 2 and 1 from column 3, or 3 from column 2 and 2 from column 3. (Remember that I can't have two elements from the same row or column.)
If I choose 1 from column 1, I can either choose 1 from column 2 and 1 from column 3, or 3 from column 2 and 4 from column 3.
Finally, if I choose 5 from column 1, I can either choose 1 from column 2 and 2 from column 3, or -1 from column 2 and 4 from column 3.
This gives me 6 products:
Next, I have to attach a sign to each product. To do this, I count the number of row swaps I need to move the 1's in the identity matrix into the same positions as the numbers in the product. I'll illustrate with two examples.
It took 2 row swaps to move the 1's into the same positions as 1, 3, and 4. Since 2 is even, the sign of is .
It took 1 row swap to move the 1's into the same positions as 7, 5, and 3. Since 1 is odd, the sign of is -1.
Continuing in this fashion, I get
Notice how ugly the computation was! While the permutation formula can be used for computations, it's easier to use row or column operations or expansion by cofactors. The main point of the permutation formula lies in the following Corollary. It says there is only one function on matrices which satisfies the three axioms for a determinant --- the determinant function is unique. Row reduction, expansion by cofactors, and the permutation formula give different ways of computing the same thing.
The permutation formula is connected to a trick for computing determinants of matrices. You may have seen this trick in other math courses, or in physics courses. I'll illustrate with the matrix in the last example.
Warning: This only works on determinants which are !
Begin by making copies of the first two columns of the matrix. Put the copies to the right of the original matrix:
Next, draw diagonal lines through the elements as shown below. Three lines down and to the right, three lines up and to the right:
Form products by multiplying the elements along each line. The products of the "down and right" lines get plus signs, and the products of the "down and left" lines get minus signs:
You can see that we got the same terms as we got with the permutation formula, with the factors and the terms in a different order.
Again, I emphasize that this trick only works on matrices which are ! You can't use it on matrices of any other size. It's not bad for determinants you're computing by hand, so feel free to use it if you wish. Don't try to use it on determinants which are , , and so on.
Corollary. Let R be a commutative ring with identity. There is a unique determinant function .
Proof. The permutation formula says
But the right side only depends on the entries of the matrix A. So is completely determined by A, and there can be only one determinant function on matrices.
We know that the determinant function defined by cofactor expansion satisfies the axioms for a determinant function. Therefore, it is the only determinant function on matrices.
This doesn't mean that you can't compute the determinant in different ways; in fact, the permutation formula gives a different way of computing determinants than cofactor expansion. To say that there's only one determinant function means that any function satisfying the determinant axioms will give the same answer as any other function satisfying the determinant axioms, for a given matrix.
Remark. Here's another way to express the theorem. Suppose denotes the determinant function on . If is alternating and linear in each row, then
In other words, a function which satisfies the first two axioms for a determinant function is a multiple of the "real" determinant function, the multiple being the value the function takes on the identity matrix. In the case of the "real" determinant function, the third axiom says , so the multiple is 1 and D is the "real" determinant function.
Copyright 2022 by Bruce Ikenaga