Determinant January 15, 2025
Multiplicative Property ¶ In this section we assume the usual definition of a determinant, from which one
can derive the following properties that we will need:
det d i a g ( d 1 , d 2 , … , d n ) = d 1 d 2 … d n , \det \mathrm{diag}(d_1, d_2, \dots, d_n)=d_1 d_2 \dots d_n\,, det diag ( d 1 , d 2 , … , d n ) = d 1 d 2 … d n ,
det A B = det A det B , \det AB = \det A \det B\,, det A B = det A det B ,
det λ A = λ n det A . \det \lambda A = \lambda^n \det A\,. det λ A = λ n det A . Now our task is to find all functions f ( A ) f(A) f ( A ) from matrices to real numbers
that satisfy the multiplicative property:
f ( A B ) = f ( A ) f ( B ) . f(AB) = f(A)f(B)\,. f ( A B ) = f ( A ) f ( B ) . We will use (4) repeatedly in the following calculation. First
we decompose the matrix A = H U A=HU A = H U using a polar decomposition, where H H H is
Hermitian and U U U is unitary. Both H H H and U U U are individually diagonalizable
(but not mutually diagonalizable, since A A A is not in general diagonalizable).
For any matrix C C C that is diagonalizable, we can write
C = P D P − 1 C=P D P^{-1} C = P D P − 1 where D D D is diagonal and P P P is invertible. Then:
f ( C ) = f ( P D P − 1 ) = f ( P ) f ( D ) f ( P − 1 ) = f ( P ) f ( P − 1 ) f ( D ) = f(C)
= f(P D P^{-1})
= f(P) f(D) f(P^{-1})
= f(P) f(P^{-1}) f(D)= f ( C ) = f ( P D P − 1 ) = f ( P ) f ( D ) f ( P − 1 ) = f ( P ) f ( P − 1 ) f ( D ) =
= f ( P P − 1 D ) = f ( D ) . = f(P P^{-1} D)
= f(D)\,. = f ( P P − 1 D ) = f ( D ) . That means that if a matrix is diagonalizable, then f f f applied to that matrix
is equal to f f f applied to the diagonal matrix.
Below we will use matrices P i P_i P i that are unit
matrices with the 1st and i i i -th rows permuted. Now we can compute, using the
fact that H H H and U U U can be decomposed into diagonal matrices D H D_H D H and D U D_U D U ,
and their product is also a diagonal matrix D = D H D U D=D_H D_U D = D H D U :
f ( A ) = f ( H U ) = f ( H ) f ( U ) = f ( D H ) f ( D U ) = f ( D H D U ) = f ( D ) = f(A)
= f(HU)
= f(H)f(U)
= f(D_H)f(D_U)
= f(D_H D_U)
= f(D) = f ( A ) = f ( H U ) = f ( H ) f ( U ) = f ( D H ) f ( D U ) = f ( D H D U ) = f ( D ) =
= f ( d i a g ( d 1 , 1 , 1 , … ) ) f ( d i a g ( 1 , d 2 , 1 , … ) ) … f ( d i a g ( 1 , … , 1 , d n ) ) = = f(\mathrm{diag}(d_1,1,1,\dots))
f(\mathrm{diag}(1,d_2,1,\dots))
\dots
f(\mathrm{diag}(1,\dots,1,d_n))= = f ( diag ( d 1 , 1 , 1 , … )) f ( diag ( 1 , d 2 , 1 , … )) … f ( diag ( 1 , … , 1 , d n )) =
= f ( P 1 d i a g ( d 1 , 1 , 1 , … ) P 1 − 1 ) f ( P 2 d i a g ( d 2 , 1 , 1 , … ) P 2 − 1 ) … = f(P_1\mathrm{diag}(d_1,1,1,\dots)P_1^{-1})
f(P_2\mathrm{diag}(d_2,1,1,\dots)P_2^{-1})
\dots = f ( P 1 diag ( d 1 , 1 , 1 , … ) P 1 − 1 ) f ( P 2 diag ( d 2 , 1 , 1 , … ) P 2 − 1 ) …
… f ( P n d i a g ( d n , 1 , 1 , … ) P n − 1 ) = \dots f(P_n\mathrm{diag}(d_n,1,1,\dots)P_n^{-1})= … f ( P n diag ( d n , 1 , 1 , … ) P n − 1 ) =
= f ( d i a g ( d 1 , 1 , 1 , … ) d i a g ( d 2 , 1 , 1 , … ) … d i a g ( d n , 1 , 1 , … ) ) = = f(\mathrm{diag}(d_1,1,1,\dots)
\mathrm{diag}(d_2,1,1,\dots)
\dots
\mathrm{diag}(d_n,1,1,\dots))= = f ( diag ( d 1 , 1 , 1 , … ) diag ( d 2 , 1 , 1 , … ) … diag ( d n , 1 , 1 , … )) =
= f ( d i a g ( d 1 d 2 … d n , 1 , 1 , … ) ) = = f(\mathrm{diag}(d_1 d_2 \dots d_n,1,1,\dots))= = f ( diag ( d 1 d 2 … d n , 1 , 1 , … )) =
= f ( d i a g ( det D , 1 , 1 , … ) ) = = f(\mathrm{diag}(\det D,1,1,\dots))= = f ( diag ( det D , 1 , 1 , … )) =
= g ( det D ) = g ( det D H D U ) = g ( det D H det D U ) = g ( det H det U ) = = g(\det D)= g(\det D_H D_U)
= g(\det D_H \det D_U)= g(\det H \det U) = = g ( det D ) = g ( det D H D U ) = g ( det D H det D U ) = g ( det H det U ) =
= g ( det H U ) = g ( det A ) , =g(\det HU) = g(\det A)\,, = g ( det H U ) = g ( det A ) , where we defined a function g ( x ) g(x) g ( x ) using
g ( x ) = f ( d i a g ( x , 1 , 1 , … ) ) . g(x) = f(\mathrm{diag}(x,1,1,\dots))\,. g ( x ) = f ( diag ( x , 1 , 1 , … )) . Let’s compute the functional equation for g ( x ) g(x) g ( x ) :
g ( x y ) = f ( d i a g ( x y , 1 , 1 , … ) ) = g(x y) = f(\mathrm{diag}(xy,1,1,\dots))= g ( x y ) = f ( diag ( x y , 1 , 1 , … )) =
= f ( d i a g ( x , 1 , 1 , … ) d i a g ( y , 1 , 1 , … ) ) = =f(\mathrm{diag}(x,1,1,\dots)\mathrm{diag}(y,1,1,\dots))= = f ( diag ( x , 1 , 1 , … ) diag ( y , 1 , 1 , … )) =
= f ( d i a g ( x , 1 , 1 , … ) ) f ( d i a g ( y , 1 , 1 , … ) ) = =f(\mathrm{diag}(x,1,1,\dots)) f(\mathrm{diag}(y,1,1,\dots))= = f ( diag ( x , 1 , 1 , … )) f ( diag ( y , 1 , 1 , … )) =
= g ( x ) g ( y ) . =g(x) g(y)\,. = g ( x ) g ( y ) . In the above computations we have only used (4) and no other
assumptions.
Let’s summarize our results. From (4) it follows that:
f ( A ) = g ( det A ) , f(A) = g(\det A)\,, f ( A ) = g ( det A ) ,
g ( x y ) = g ( x ) g ( y ) , g(xy) = g(x)g(y)\,, g ( x y ) = g ( x ) g ( y ) , for numbers x x x , y y y .
The equation (22) can be solved in several different ways.
Approach 1: Assuming that f ( A ) f(A) f ( A ) (and thus g ( x ) g(x) g ( x ) ) is measurable, we show below
that the only solutions are g ( x ) = 0 g(x)=0 g ( x ) = 0 , g ( x ) = ∣ s i g n ( x ) ∣ g(x)=|\mathrm{sign}(x)| g ( x ) = ∣ sign ( x ) ∣ , g ( x ) = ∣ x ∣ s g(x)=|x|^s g ( x ) = ∣ x ∣ s
and g ( x ) = s i g n ( x ) ∣ x ∣ s g(x)=\mathrm{sign}(x)|x|^s g ( x ) = sign ( x ) ∣ x ∣ s . Consequently the only nonzero solution to
(4) is a power of a determinant (optionally multiplied with
the signum function). If f ( A ) f(A) f ( A ) is not measurable then we also get highly
pathological solutions (see the section below). The above is thus the most
general solution.
Approach 2: Instead of assuming measurability, we can instead
assume homogeneity:
f ( λ A ) = λ n f ( A ) , f(\lambda A) = \lambda^n f(A)\,, f ( λ A ) = λ n f ( A ) , where n n n is the (integer) dimension of the matrix A A A .
Using (21) and (22) we get:
f ( λ A ) = g ( det λ A ) = g ( λ n det A ) = g ( λ n ) g ( det A ) = g ( λ n ) f ( A ) = f(\lambda A) = g(\det \lambda A)
=g(\lambda^n \det A)
=g(\lambda^n) g(\det A)
=g(\lambda^n) f(A)= f ( λ A ) = g ( det λ A ) = g ( λ n det A ) = g ( λ n ) g ( det A ) = g ( λ n ) f ( A ) =
= g ( λ ) n f ( A ) = f ( g ( λ ) A ) . =g(\lambda)^n f(A)
=f(g(\lambda) A)\,. = g ( λ ) n f ( A ) = f ( g ( λ ) A ) . Comparing the LHS and RHS we can see that g ( λ ) = λ g(\lambda)=\lambda g ( λ ) = λ , or:
and from (21) we get:
f ( A ) = det A . f(A) = \det A\,. f ( A ) = det A . Summary:
Assuming equations (4) and (23) we obtained
(27) as the only nonzero solution. That means that we
can define a determinant using (4) and (23) and
no other assumptions.
Alternatively, if we only assume (4) and that f ( A ) f(A) f ( A ) is
measurable, then the only nonzero solution is a power of a determinant
(optionally multiplied with the signum function).
Cauchy’s Additive Functional Equation ¶ The Cauchy’s functional equation (Wikipedia (2025) ) usually means
the additive equation of this form:
f ( x + y ) = f ( x ) + f ( y ) . f(x+y) = f(x) + f(y)\,. f ( x + y ) = f ( x ) + f ( y ) . There are other related equations, such as the Cauchy’s multiplicative
functional equation in the next section.
For integer n n n we get:
f ( n x ) = f ( x + x + ⋯ + x ⏟ n ) = f ( x ) + f ( x ) + ⋯ + f ( x ) ⏟ n = n f ( x ) . f(nx)
= f(\underbrace{x+x+\dots+x}_{n})
= \underbrace{f(x)+f(x)+\dots+f(x)}_{n}
= nf(x)\,. f ( n x ) = f ( n x + x + ⋯ + x ) = n f ( x ) + f ( x ) + ⋯ + f ( x ) = n f ( x ) . By substituting y = n x y=nx y = n x we get f ( y ) = n f ( y / n ) f(y)=nf(y/n) f ( y ) = n f ( y / n ) , from which
f ( y / n ) = ( 1 / n ) f ( y ) f(y/n)=(1/n)f(y) f ( y / n ) = ( 1/ n ) f ( y ) . Using these two relations for multiplication and division
by an integer we obtain for any rational number q = p q q={p\over q} q = q p , where p p p , q q q
are integers:
f ( q x ) = f ( p q x ) = p f ( x q ) = p q f ( x ) = q f ( x ) . f(qx)
=f\left({p \over q}x\right)
=pf\left({x\over q}\right)
={p\over q}f(x)
=q f(x)
\,. f ( q x ) = f ( q p x ) = p f ( q x ) = q p f ( x ) = q f ( x ) . We also have to prove it for p ≤ 0 p\le 0 p ≤ 0 which we can using f ( 0 ) = f ( 0 ) + f ( 0 ) f(0)=f(0)+f(0) f ( 0 ) = f ( 0 ) + f ( 0 ) which
implies f ( 0 ) = 0 f(0)=0 f ( 0 ) = 0 and using 0 = f ( 0 ) = f ( x − x ) = f ( x ) + f ( − x ) 0=f(0)=f(x-x)=f(x)+f(-x) 0 = f ( 0 ) = f ( x − x ) = f ( x ) + f ( − x ) from which follows
f ( − x ) = − f ( x ) f(-x)=-f(x) f ( − x ) = − f ( x ) . Setting x = 1 x=1 x = 1 we get for all rational q q q :
f ( q ) = q f ( 1 ) = c q . f(q) = q f(1) = c q\,. f ( q ) = q f ( 1 ) = c q . where c = f ( 1 ) c=f(1) c = f ( 1 ) is any real constant. If we assume that f ( x ) f(x) f ( x ) is continuous,
then this implies that f ( x ) = c x f(x)=c x f ( x ) = c x for all real x x x , since the rational numbers
are dense in real numbers. It turns out that one can relax the continuity
requirement to only require that f ( x ) f(x) f ( x ) is measurable. Since f ( x ) = c x f(x)=c x f ( x ) = c x is
linear, we get:
A linear function f ( x ) = a x + b f(x)=ax+b f ( x ) = a x + b can be substituted into (28) and one
obtains that b = 0 b=0 b = 0 . So this theorem is a simple way to remember the solutions
to the Cauchy’s additive functional equation.
To prove this theorem, we introduce g ( x ) = f ( x ) − f ( 1 ) x g(x)=f(x)-f(1)x g ( x ) = f ( x ) − f ( 1 ) x and prove that g ( x ) = 0 g(x)=0 g ( x ) = 0 ,
from which it follows that f ( x ) = f ( 1 ) x = c x f(x)=f(1)x=c x f ( x ) = f ( 1 ) x = c x . We have:
g ( x + y ) = f ( x + y ) − f ( 1 ) ( x + y ) = f ( x ) − f ( 1 ) x + f ( y ) − f ( 1 ) y = g ( x ) + g ( y ) , g(x+y)=f(x+y)-f(1)(x+y)=f(x)-f(1)x + f(y)-f(1)y = g(x)+g(y)\,, g ( x + y ) = f ( x + y ) − f ( 1 ) ( x + y ) = f ( x ) − f ( 1 ) x + f ( y ) − f ( 1 ) y = g ( x ) + g ( y ) , so g ( x ) g(x) g ( x ) is also additive. For rational q q q we have:
g ( q ) = f ( q ) − f ( 1 ) q = q f ( 1 ) − f ( 1 ) q = 0 . g(q)=f(q)-f(1)q=q f(1) - f(1)q = 0\,. g ( q ) = f ( q ) − f ( 1 ) q = q f ( 1 ) − f ( 1 ) q = 0 . Also we get:
g ( x + q ) = g ( x ) + g ( q ) = g ( x ) , g(x+q)=g(x)+g(q)=g(x)\,, g ( x + q ) = g ( x ) + g ( q ) = g ( x ) , so g ( x ) g(x) g ( x ) is periodic with a period q q q . It can be shown from this periodicity
and function g ( x ) g(x) g ( x ) being measurable that g ( x ) g(x) g ( x ) is essentially bounded. Now we
can compute the following integral which is exists and is finite:
∫ 0 1 g ( x ) d x = ∫ 0 1 g ( x + y ) d x = ∫ 0 1 g ( x ) d x + ∫ 0 1 g ( y ) d x = ∫ 0 1 g ( x ) d x + g ( y ) , \int_0^1 g(x) dx
=\int_0^1 g(x+y) dx
=\int_0^1 g(x) dx +\int_0^1 g(y) dx
=\int_0^1 g(x) dx +g(y)\,, ∫ 0 1 g ( x ) d x = ∫ 0 1 g ( x + y ) d x = ∫ 0 1 g ( x ) d x + ∫ 0 1 g ( y ) d x = ∫ 0 1 g ( x ) d x + g ( y ) , so g ( y ) = 0 g(y)=0 g ( y ) = 0 for all real y y y . Here we have used the dominated convergence
theorem for g ( x ) = g ( x + q n ) g(x)=g(x+q_n) g ( x ) = g ( x + q n ) where a sequence of rational q n q_n q n is converging
towards the real y y y .
Cauchy’s Multiplicative Functional Equation ¶ The Cauchy’s multiplicative functional equation is:
f ( x y ) = f ( x ) f ( y ) . f(xy) = f(x)f(y)\,. f ( x y ) = f ( x ) f ( y ) . There are only 3 cases that can happen:
f ( x ) = 0 . f(x) = 0\,. f ( x ) = 0 .
f ( x ) = 1 . f(x) = 1\,. f ( x ) = 1 .
f ( x ) = { h ( x ) if x > 0 , 0 if x = 0 , ± h ( ∣ x ∣ ) if x < 0 . f(x) = \begin{cases}
h(x) & \text{if } x > 0\,,\\
0 & \text{if } x = 0\,,\\
\pm h(|x|) & \text{if } x < 0\,.\\
\end{cases} f ( x ) = ⎩ ⎨ ⎧ h ( x ) 0 ± h ( ∣ x ∣ ) if x > 0 , if x = 0 , if x < 0 .
In the last case the function h ( x ) h(x) h ( x ) is positive h ( x ) > 0 h(x) > 0 h ( x ) > 0 and defined only
for x > 0 x>0 x > 0 . The x < 0 x < 0 x < 0 case has two options for even and odd f ( x ) f(x) f ( x ) . We can
equivalently write the two cases explicitly as follows:
Even: f ( x ) = { 0 if x = 0 , h ( ∣ x ∣ ) if x ≠ 0 . f(x) = \begin{cases}
0 & \text{if } x = 0\,,\\
h(|x|) & \text{if } x \ne 0\,.\\
\end{cases} f ( x ) = { 0 h ( ∣ x ∣ ) if x = 0 , if x = 0 .
Odd: f ( x ) = { 0 if x = 0 , s i g n ( x ) h ( ∣ x ∣ ) if x ≠ 0 . f(x) = \begin{cases}
0 & \text{if } x = 0\,,\\
\mathrm{sign}(x) h(|x|) & \text{if } x \ne 0\,.\\
\end{cases} f ( x ) = { 0 sign ( x ) h ( ∣ x ∣ ) if x = 0 , if x = 0 .
As a special case for h ( x ) = 1 h(x)=1 h ( x ) = 1 we get:
Even: f ( x ) = ∣ s i g n ( x ) ∣ = { 0 if x = 0 , 1 if x ≠ 0 . f(x) = |\mathrm{sign}(x)| = \begin{cases}
0 & \text{if } x = 0\,,\\
1 & \text{if } x \ne 0\,.
\end{cases} f ( x ) = ∣ sign ( x ) ∣ = { 0 1 if x = 0 , if x = 0 .
Odd: f ( x ) = s i g n ( x ) = { 0 if x = 0 , 1 if x > 0 , − 1 if x < 0 . f(x) = \mathrm{sign}(x) = \begin{cases}
0 & \text{if } x = 0\,,\\
1 & \text{if } x > 0\,,\\
-1 & \text{if } x < 0\,.
\end{cases} f ( x ) = sign ( x ) = ⎩ ⎨ ⎧ 0 1 − 1 if x = 0 , if x > 0 , if x < 0 .
We can then use the substitution g ( x ) = log f ( e x ) g(x)=\log f(e^x) g ( x ) = log f ( e x ) to convert
the multiplicative equation into an additive equation from the previous
section:
g ( x + y ) = log f ( e x + y ) = log f ( e x e y ) = log ( f ( e x ) f ( e y ) ) = g(x+y)
= \log f(e^{x+y})
= \log f(e^x e^y)
= \log(f(e^x) f(e^y))= g ( x + y ) = log f ( e x + y ) = log f ( e x e y ) = log ( f ( e x ) f ( e y )) =
= log f ( e x ) + log f ( e y ) = g ( x ) + g ( y ) . = \log f(e^x) + \log f(e^y)
= g(x) + g(y)\,. = log f ( e x ) + log f ( e y ) = g ( x ) + g ( y ) . We find the solution using the previous section, and then
we compute h ( x ) h(x) h ( x ) using h ( x ) = exp ( g ( log x ) ) h(x)=\exp(g(\log x)) h ( x ) = exp ( g ( log x )) , which satisfies:
h ( x y ) = exp ( g ( log ( x y ) ) ) = exp ( g ( log x + log y ) ) = exp ( g ( log x ) + g ( log y ) ) = h(xy)
=\exp(g(\log(xy)))
=\exp(g(\log x + \log y))
=\exp(g(\log x) + g(\log y))= h ( x y ) = exp ( g ( log ( x y ))) = exp ( g ( log x + log y )) = exp ( g ( log x ) + g ( log y )) =
= exp ( g ( log x ) ) exp ( g ( log y ) ) = h ( x ) h ( y ) . =\exp(g(\log x))\exp(g(\log y))
=h(x)h(y)\,. = exp ( g ( log x )) exp ( g ( log y )) = h ( x ) h ( y ) . For measurable g ( x ) g(x) g ( x ) the solution to the additive equation is g ( x ) = c x g(x) = cx g ( x ) = c x , so
we assume f ( x ) f(x) f ( x ) is measurable (which implies g ( x ) g(x) g ( x ) is measurable) and
we get h ( x ) = exp ( c log x ) = exp ( log ( x c ) ) = x c h(x)=\exp(c\log x)=\exp(\log(x^c))=x^c h ( x ) = exp ( c log x ) = exp ( log ( x c )) = x c and the four solutions are
f ( x ) = 0 . f(x) = 0\,. f ( x ) = 0 .
f ( x ) = 1 . f(x) = 1\,. f ( x ) = 1 .
f ( x ) = { 0 if x = 0 , h ( ∣ x ∣ ) if x ≠ 0 , = { 0 if x = 0 , ∣ x ∣ c if x ≠ 0 . f(x) = \begin{cases}
0 & \text{if } x = 0\,,\\
h(|x|) & \text{if } x \ne 0\,,\\
\end{cases}
= \begin{cases}
0 & \text{if } x = 0\,,\\
|x|^c & \text{if } x \ne 0\,.\\
\end{cases} f ( x ) = { 0 h ( ∣ x ∣ ) if x = 0 , if x = 0 , = { 0 ∣ x ∣ c if x = 0 , if x = 0 .
f ( x ) = { 0 if x = 0 , s i g n ( x ) h ( ∣ x ∣ ) if x ≠ 0 , = { 0 if x = 0 , s i g n ( x ) ∣ x ∣ c if x ≠ 0 . f(x) = \begin{cases}
0 & \text{if } x = 0\,,\\
\mathrm{sign}(x) h(|x|) & \text{if } x \ne 0\,,\\
\end{cases}
= \begin{cases}
0 & \text{if } x = 0\,,\\
\mathrm{sign}(x) |x|^c & \text{if } x \ne 0\,.\\
\end{cases} f ( x ) = { 0 sign ( x ) h ( ∣ x ∣ ) if x = 0 , if x = 0 , = { 0 sign ( x ) ∣ x ∣ c if x = 0 , if x = 0 .
The solutions can be equivalently written as:
f ( x ) = 0 . f(x)=0\,. f ( x ) = 0 .
f ( x ) = ∣ s i g n ( x ) ∣ = { 0 if x = 0 , 1 if x ≠ 0 . f(x) = |\mathrm{sign}(x)| = \begin{cases}
0 & \text{if } x = 0\,,\\
1 & \text{if } x \ne 0\,.
\end{cases} f ( x ) = ∣ sign ( x ) ∣ = { 0 1 if x = 0 , if x = 0 .
f ( x ) = ∣ x ∣ c f(x)=|x|^c f ( x ) = ∣ x ∣ c , with f ( 0 ) = 0 f(0)=0 f ( 0 ) = 0 (needed for c < 0 ) c < 0) c < 0 ) .
f ( x ) = s i g n ( x ) ∣ x ∣ c f(x)=\mathrm{sign}(x)|x|^c f ( x ) = sign ( x ) ∣ x ∣ c , with f ( 0 ) = 0 f(0)=0 f ( 0 ) = 0 (needed for c < 0 ) c < 0) c < 0 ) .
The solution f ( x ) = 1 f(x)=1 f ( x ) = 1 is already included in ∣ x ∣ c |x|^c ∣ x ∣ c for c = 0 c=0 c = 0 .
If c c c is an integer, then the last two solutions can always be written (for
both even and odd integers) as f ( x ) = x c f(x)=x^c f ( x ) = x c and f ( x ) = s i g n ( x ) x c f(x)=\mathrm{sign}(x) x^c f ( x ) = sign ( x ) x c .
Some examples are: 0, 1, s i g n ( x ) \mathrm{sign}(x) sign ( x ) , ∣ s i g n ( x ) ∣ |\mathrm{sign}(x)| ∣ sign ( x ) ∣ , x x x ,
∣ x ∣ |x| ∣ x ∣ , x 2 x^2 x 2 , s i g n ( x ) x 2 \mathrm{sign}(x) x^2 sign ( x ) x 2 , x 3 x^3 x 3 , ∣ x ∣ 3 |x|^3 ∣ x ∣ 3 , 1 x 1\over x x 1 , 1 ∣ x ∣ 1\over
|x| ∣ x ∣ 1 ,1 x 2 1\over x^2 x 2 1 ,s i g n ( x ) x 2 = ∣ x ∣ x 3 {\mathrm{sign}(x)\over x^2}={|x|\over x^3} x 2 sign ( x ) = x 3 ∣ x ∣ , 1 x 3 1\over x^3 x 3 1 ,
∣ x ∣ \sqrt{|x|} ∣ x ∣ , 1 ∣ x ∣ 3 1\over\sqrt{|x|^3} ∣ x ∣ 3 1 , etc.
The function h ( x ) h(x) h ( x ) is only defined for positive x x x and that is all we need.
But if we are curious how the formula evaluates for x ≤ 0 x \le 0 x ≤ 0 , we get the
following.
For c > 0 c>0 c > 0 and x = 0 + x=0^+ x = 0 + we get:
h ( x ) = exp ( c log 0 + ) = exp ( − c ∞ ) = 0 . h(x)=\exp(c \log 0^+)=\exp(-c\,\infty)=0\,. h ( x ) = exp ( c log 0 + ) = exp ( − c ∞ ) = 0 . For x < 0 x<0 x < 0 we get:
h ( x ) = exp ( c ( log ∣ x ∣ + i arg x ) ) = exp ( c ( log ( − x ) + i π ) ) = h(x)
=\exp(c (\log |x| + i \arg x))
=\exp(c (\log(-x) + i \pi))= h ( x ) = exp ( c ( log ∣ x ∣ + i arg x )) = exp ( c ( log ( − x ) + iπ )) =
= exp ( log ( ( − x ) c ) ) exp ( i c π ) = ( − x ) c exp ( i c π ) . =\exp(\log((-x)^c))\exp(i c \pi)
=(-x)^c \exp(i c \pi)\,. = exp ( log (( − x ) c )) exp ( i c π ) = ( − x ) c exp ( i c π ) . For an even integer c c c we have exp ( i c π ) = 1 \exp(ic\pi)=1 exp ( i c π ) = 1 and h ( x ) = ( − x ) c = x c h(x)=(-x)^c=x^c h ( x ) = ( − x ) c = x c , for an
odd integer c c c we have exp ( i c π ) = − 1 \exp(ic\pi)=-1 exp ( i c π ) = − 1 and h ( x ) = − ( − x ) c = x c h(x)=-(-x)^c=x^c h ( x ) = − ( − x ) c = x c and for
non-integer c c c we get some complex number. For example, for c = 1 c=1 c = 1 (which is an
odd integer) we get:
h ( x ) = ( − x ) ( − 1 ) = x . h(x) = (-x) (-1) = x\,. h ( x ) = ( − x ) ( − 1 ) = x . Due to the discontinuity at x = 0 x=0 x = 0 , we can stitch solutions there, and extend
h ( x ) h(x) h ( x ) as even or odd functions.
Geometric Definition ¶ The way to define a determinant using area/volume from geometry is as follows:
V ( e 1 , e 2 ) = 1 , V(\mathbf{e}_1,\mathbf{e}_2) = 1\,, V ( e 1 , e 2 ) = 1 ,
V ( u , u ) = 0 , V(\mathbf{u},\mathbf{u}) = 0\,, V ( u , u ) = 0 , for c > 0 c>0 c > 0 :
V ( c u , w ) = c V ( u , w ) = V ( u , c w ) , V(c\mathbf{u}, \mathbf{w})
= c V(\mathbf{u}, \mathbf{w}) = V(\mathbf{u},
c\mathbf{w})\,, V ( c u , w ) = c V ( u , w ) = V ( u , c w ) ,
V ( u + v , w ) = V ( u , w ) + V ( v , w ) , V(\mathbf{u}+\mathbf{v}, \mathbf{w})
= V(\mathbf{u}, \mathbf{w}) + V(\mathbf{v}, \mathbf{w})\,, V ( u + v , w ) = V ( u , w ) + V ( v , w ) ,
V ( w , u + v ) = V ( w , u ) + V ( w , v ) . V(\mathbf{w}, \mathbf{u}+\mathbf{v})
= V(\mathbf{w}, \mathbf{u}) + V(\mathbf{w}, \mathbf{v})\,. V ( w , u + v ) = V ( w , u ) + V ( w , v ) . All these formulas can be “derived” from intuitive properties of areas in 2D.
Another equivalent set of requirements is:
V ( e 1 , e 2 ) = 1 , V(\mathbf{e}_1,\mathbf{e}_2) = 1\,, V ( e 1 , e 2 ) = 1 ,
V ( c u , w ) = c V ( u , w ) = V ( u , c w ) , V(c\mathbf{u}, \mathbf{w})
= c V(\mathbf{u}, \mathbf{w}) = V(\mathbf{u},
c\mathbf{w})\,, V ( c u , w ) = c V ( u , w ) = V ( u , c w ) ,
V ( u + v , v ) = V ( u , v ) = V ( u , u + v ) . V(\mathbf{u}+\mathbf{v}, \mathbf{v})
= V(\mathbf{u}, \mathbf{v})
= V(\mathbf{u}, \mathbf{u}+\mathbf{v})\,. V ( u + v , v ) = V ( u , v ) = V ( u , u + v ) . Let’s use the first set of requirements below. We get:
0 = V ( v − v , w ) = V ( v , w ) + V ( − v , w ) , 0 = V(\mathbf{v}-\mathbf{v}, \mathbf{w}) = V(\mathbf{v}, \mathbf{w}) + V(-\mathbf{v}, \mathbf{w})\,, 0 = V ( v − v , w ) = V ( v , w ) + V ( − v , w ) , from which it follows:
V ( − v , w ) = − V ( v , w ) . V(-\mathbf{v}, \mathbf{w}) = -V(\mathbf{v}, \mathbf{w})\,. V ( − v , w ) = − V ( v , w ) . Now we derive antisymmetry:
0 = V ( u + v , u + v ) = 0
= V(\mathbf{u}+\mathbf{v}, \mathbf{u}+\mathbf{v}) = 0 = V ( u + v , u + v ) =
= V ( u , u ) + V ( u , v ) + V ( v , u ) + V ( v , v ) = V ( u , v ) + V ( v , u ) = V(\mathbf{u},\mathbf{u}) + V(\mathbf{u},\mathbf{v}) + V(\mathbf{v},\mathbf{u}) + V(\mathbf{v},\mathbf{v}) = V(\mathbf{u},\mathbf{v}) + V(\mathbf{v},\mathbf{u}) = V ( u , u ) + V ( u , v ) + V ( v , u ) + V ( v , v ) = V ( u , v ) + V ( v , u ) and we get:
V ( v , u ) = − V ( u , v ) . V(\mathbf{v},\mathbf{u}) = -V(\mathbf{u},\mathbf{v})\,. V ( v , u ) = − V ( u , v ) . Now we derive a formula for V V V :
V ( u , v ) = V ( u 1 e 1 + u 2 e 2 , v 1 e 1 + v 2 e 2 ) = V(\mathbf{u},\mathbf{v})
= V(u_1\mathbf{e}_1+u_2\mathbf{e}_2, v_1\mathbf{e}_1+v_2\mathbf{e}_2) = V ( u , v ) = V ( u 1 e 1 + u 2 e 2 , v 1 e 1 + v 2 e 2 ) =
= u 1 v 2 V ( e 1 , e 2 ) + u 2 v 1 V ( e 2 , e 1 ) = ( u 1 v 2 − u 2 v 1 ) V ( e 1 , e 2 ) = u 1 v 2 − u 2 v 1 . = u_1v_2V(\mathbf{e}_1,\mathbf{e}_2) + u_2v_1V(\mathbf{e}_2,\mathbf{e}_1)
= (u_1v_2-u_2v_1)V(\mathbf{e}_1,\mathbf{e}_2)
= u_1v_2-u_2v_1\,. = u 1 v 2 V ( e 1 , e 2 ) + u 2 v 1 V ( e 2 , e 1 ) = ( u 1 v 2 − u 2 v 1 ) V ( e 1 , e 2 ) = u 1 v 2 − u 2 v 1 . This proves both existence and uniqueness.
Coordinate Definition ¶ Determinant of a 2x2 matrix ( A ) i j = a i j (\mathbf{A})_{ij}=a_{ij} ( A ) ij = a ij is defined as:
det A = ϵ i j a 1 i a 2 j = 1 2 ! ϵ i j ϵ k l a k i a l j \det \mathbf{A} = \epsilon^{ij} a_{1i}a_{2j}
= {1\over 2!}\epsilon^{ij}\epsilon^{kl} a_{ki}a_{lj} det A = ϵ ij a 1 i a 2 j = 2 ! 1 ϵ ij ϵ k l a ki a l j and for a 3x3 matrix ( A ) i j = a i j (\mathbf{A})_{ij}=a_{ij} ( A ) ij = a ij :
det A = ϵ i j k a 1 i a 2 j a 3 k = 1 3 ! ϵ i j k ϵ l m n a l i a m j a n k \det \mathbf{A} = \epsilon^{ijk} a_{1i}a_{2j}a_{3k}
= {1\over 3!} \epsilon^{ijk}\epsilon^{lmn} a_{li}a_{mj}a_{nk} det A = ϵ ijk a 1 i a 2 j a 3 k = 3 ! 1 ϵ ijk ϵ l mn a l i a mj a nk And so on in any dimension. The symbol ϵ i j k \epsilon^{ijk} ϵ ijk is a Levi-Civita
symbol. All the properties of determinants are encoded in the Levi-Civita
symbol, so we need to understand it well, but the rest are just usual tensor
manipulations.
For the 2x2 matrix we can compute the second formula on the RHS as follows:
det A = 1 2 ! ϵ i j ϵ k l a k i a l j = 1 2 ! det ( δ i k δ i l δ j k δ j l ) a k i a l j = \det \mathbf{A}
= {1\over 2!}\epsilon^{ij}\epsilon^{kl} a_{ki}a_{lj}
= {1\over 2!}\det\begin{pmatrix} \delta^{ik} & \delta^{il} \\
\delta^{jk} & \delta^{jl} \end{pmatrix} a_{ki}a_{lj} = det A = 2 ! 1 ϵ ij ϵ k l a ki a l j = 2 ! 1 det ( δ ik δ jk δ i l δ j l ) a ki a l j =
= 1 2 ! ( δ i k δ j l − δ i l δ j k ) a k i a l j = 1 2 ! ( a i i a j j − a j i a i j ) = = {1\over 2!}( \delta^{ik}\delta^{jl} - \delta^{il} \delta^{jk})
a_{ki}a_{lj}
= {1\over 2!}( a_{ii}a_{jj} - a_{ji}a_{ij}) = = 2 ! 1 ( δ ik δ j l − δ i l δ jk ) a ki a l j = 2 ! 1 ( a ii a jj − a ji a ij ) =
= 1 2 ! ( ( a 11 + a 22 ) ( a 11 + a 22 ) − ( a 11 a 11 + a 12 a 21 + a 21 a 12 + a 22 a 22 ) ) = = {1\over 2!}( (a_{11}+a_{22})(a_{11}+a_{22}) - (a_{11}a_{11}+a_{12}a_{21}+a_{21}a_{12}+a_{22}a_{22})) = = 2 ! 1 (( a 11 + a 22 ) ( a 11 + a 22 ) − ( a 11 a 11 + a 12 a 21 + a 21 a 12 + a 22 a 22 )) =
= 1 2 ! ( 2 a 11 a 22 − 2 a 21 a 12 ) = a 11 a 22 − a 21 a 12 . = {1\over 2!}( 2a_{11}a_{22} - 2a_{21}a_{12})
= a_{11}a_{22} - a_{21}a_{12}\,. = 2 ! 1 ( 2 a 11 a 22 − 2 a 21 a 12 ) = a 11 a 22 − a 21 a 12 . The first formula on the RHS is just:
det A = ϵ i j a 1 i a 2 j = a 11 a 22 − a 12 a 21 . \det \mathbf{A}
= \epsilon^{ij} a_{1i}a_{2j}
= a_{11}a_{22}-a_{12}a_{21}\,. det A = ϵ ij a 1 i a 2 j = a 11 a 22 − a 12 a 21 . Both results are equivalent. It works the same way in any dimension, the
factorial factor corrects the overcounting.
Let us now derive the multiplicative formula for determinants:
det A B = det A det B . \det\mathbf{AB} = \det\mathbf{A}\det\mathbf{B}\,. det AB = det A det B . Let’s show it for 2x2 matrices.
The LHS is:
det A B = ϵ i j ( A B ) 1 i ( A B ) 2 j = ϵ i j ( a 1 k b k i ) ( a 2 l b l j ) . \det\mathbf{AB}
= \epsilon^{ij} (\mathbf{AB})_{1i}(\mathbf{AB})_{2j}
= \epsilon^{ij} (a_1{}^k b_{ki}) (a_2{}^l b_{lj})\,. det AB = ϵ ij ( AB ) 1 i ( AB ) 2 j = ϵ ij ( a 1 k b ki ) ( a 2 l b l j ) . The RHS is:
det A det B = ϵ i j a 1 i a 2 j ϵ k l b 1 k b 2 l = \det\mathbf{A}\det\mathbf{B}
= \epsilon^{ij} a_{1i}a_{2j} \epsilon^{kl} b_{1k}b_{2l}= det A det B = ϵ ij a 1 i a 2 j ϵ k l b 1 k b 2 l =
= ( δ i k δ j l − δ i l δ j k ) a 1 i a 2 j b 1 k b 2 l = = (\delta^{ik}\delta^{jl} - \delta^{il}\delta^{jk}) a_{1i}a_{2j} b_{1k}b_{2l}= = ( δ ik δ j l − δ i l δ jk ) a 1 i a 2 j b 1 k b 2 l =
= a 1 i a 2 j b 1 i b 2 j − a 1 i a 2 j b 1 j b 2 i = = a_{1i}a_{2j} b_1{}^i b_2{}^j - a_{1i}a_{2j} b_1{}^j b_2{}^i = = a 1 i a 2 j b 1 i b 2 j − a 1 i a 2 j b 1 j b 2 i =
= a 1 i a 2 j ( b 1 i b 2 j − b 1 j b 2 i ) = = a_{1i}a_{2j} (b_1{}^i b_2{}^j - b_1{}^j b_2{}^i) = = a 1 i a 2 j ( b 1 i b 2 j − b 1 j b 2 i ) =
= ϵ k l a 1 i a 2 j b k i b l j = = \epsilon^{kl}a_{1i}a_{2j} b_k{}^i b_l{}^j = = ϵ k l a 1 i a 2 j b k i b l j =
= ϵ k l a 1 i b k i a 2 j b l j = = \epsilon^{kl}a_{1i} b_k{}^i a_{2j} b_l{}^j = = ϵ k l a 1 i b k i a 2 j b l j =
= ϵ i j a 1 k b i k a 2 l b j l = = \epsilon^{ij}a_{1k} b_i{}^k a_{2l} b_j{}^l = = ϵ ij a 1 k b i k a 2 l b j l =
= ϵ i j a 1 k b i k a 2 l b j l = = \epsilon^{ij}a_1{}^k b_{ik} a_2{}^l b_{jl} = = ϵ ij a 1 k b ik a 2 l b j l =
= det A B T . =\det\mathbf{AB}^T\,. = det AB T . Now we need to prove det A T = det A \det\mathbf{A}^T = \det\mathbf{A} det A T = det A . We can do that using
the factorial formula by relabeling indices as follows:
det A T = 1 2 ! ϵ i j ϵ k l a i k a j l = 1 2 ! ϵ k l ϵ i j a k i a l j = 1 2 ! ϵ i j ϵ k l a k i a l j = det A . \det\mathbf{A}^T
= {1\over 2!}\epsilon^{ij}\epsilon^{kl} a_{ik}a_{jl}
= {1\over 2!}\epsilon^{kl}\epsilon^{ij} a_{ki}a_{lj}
= {1\over 2!}\epsilon^{ij}\epsilon^{kl} a_{ki}a_{lj}
= \det\mathbf{A}\,. det A T = 2 ! 1 ϵ ij ϵ k l a ik a j l = 2 ! 1 ϵ k l ϵ ij a ki a l j = 2 ! 1 ϵ ij ϵ k l a ki a l j = det A . We have thus shown:
det A B = det A det B T = det A det B . \det\mathbf{AB}
=\det\mathbf{A}\det\mathbf{B}^T
=\det\mathbf{A}\det\mathbf{B}\,. det AB = det A det B T = det A det B . Alternative Proof ¶ First we need the following identity:
det ( A ) ϵ i j = ϵ k l a 1 k a 2 l ϵ i j = det ( δ i k δ i l δ j k δ j l ) a 1 k a 2 l = det ( a 1 i a 2 i a 1 j a 2 j ) = ϵ k l a k i a l j . \det({\mathbf A}) \epsilon_{ij}
= \epsilon^{kl} a_{1k}a_{2l} \epsilon_{ij}
= \det\begin{pmatrix}
\delta_i{}^{k} & \delta_i{}^{l} \\
\delta_j{}^{k} & \delta_j{}^{l} \end{pmatrix} a_{1k}a_{2l}
= \det\begin{pmatrix}
a_{1i} & a_{2i} \\
a_{1j} & a_{2j} \end{pmatrix}
= \epsilon^{kl} a_{ki} a_{lj}\,. det ( A ) ϵ ij = ϵ k l a 1 k a 2 l ϵ ij = det ( δ i k δ j k δ i l δ j l ) a 1 k a 2 l = det ( a 1 i a 1 j a 2 i a 2 j ) = ϵ k l a ki a l j . The final step above follows from using the definition of a determinant as
follows:
det ( X ) = det ( a 1 i a 2 i a 1 j a 2 j ) = ϵ k l X 1 k X 2 l . \det(\mathbf{X})
= \det\begin{pmatrix}
a_{1i} & a_{2i} \\
a_{1j} & a_{2j} \end{pmatrix}
=\epsilon^{kl} \mathbf{X}_{1k} \mathbf{X}_{2l}\,. det ( X ) = det ( a 1 i a 1 j a 2 i a 2 j ) = ϵ k l X 1 k X 2 l . Now X 11 = a 1 i \mathbf{X}_{11}=a_{1i} X 11 = a 1 i and X 12 = a 2 i \mathbf{X}_{12}=a_{2i} X 12 = a 2 i , so
X 1 k = a k i \mathbf{X}_{1k}=a_{ki} X 1 k = a ki . Similarly X 2 l = a l j \mathbf{X}_{2l}=a_{lj} X 2 l = a l j , so we get:
= ϵ k l a k i a l j . =\epsilon^{kl} a_{ki} a_{lj}\,. = ϵ k l a ki a l j . Alternatively, one can simply expand the determinant as follows:
det ( a 1 i a 2 i a 1 j a 2 j ) = a 1 i a 2 j − a 2 i a 1 j = ϵ k l a k i a l j . \det\begin{pmatrix}
a_{1i} & a_{2i} \\
a_{1j} & a_{2j} \end{pmatrix}
= a_{1i} a_{2j} - a_{2i} a_{1j}
=\epsilon^{kl} a_{ki} a_{lj}\,. det ( a 1 i a 1 j a 2 i a 2 j ) = a 1 i a 2 j − a 2 i a 1 j = ϵ k l a ki a l j . Contracting both sides with ϵ i j \epsilon^{ij} ϵ ij we get:
det ( A ) ϵ i j ϵ i j = ϵ i j ϵ k l a k i a l j , \det({\mathbf A}) \epsilon_{ij} \epsilon^{ij}
=\epsilon^{ij}\epsilon^{kl} a_{ki} a_{lj}\,, det ( A ) ϵ ij ϵ ij = ϵ ij ϵ k l a ki a l j ,
det ( A ) 2 ! = ϵ i j ϵ k l a k i a l j , \det({\mathbf A})\, 2!
=\epsilon^{ij}\epsilon^{kl} a_{ki} a_{lj}\,, det ( A ) 2 ! = ϵ ij ϵ k l a ki a l j ,
det ( A ) = 1 2 ! ϵ i j ϵ k l a k i a l j . \det({\mathbf A})
={1\over2!}\epsilon^{ij}\epsilon^{kl} a_{ki} a_{lj}\,. det ( A ) = 2 ! 1 ϵ ij ϵ k l a ki a l j . Note that ϵ k l a k i a l j = ϵ k l a i k a j l \epsilon^{kl} a_{ki} a_{lj} = \epsilon^{kl} a_{ik} a_{jl} ϵ k l a ki a l j = ϵ k l a ik a j l due to
the transposition property derived using the last equation in the previous
section:
det ( A ) ϵ i j = ϵ k l a k i a l j = det ( A T ) ϵ i j = ϵ k l a i k a j l . \det({\mathbf A}) \epsilon_{ij}
= \epsilon^{kl} a_{ki} a_{lj}
= \det({\mathbf A}^T) \epsilon_{ij}
= \epsilon^{kl} a_{ik} a_{jl}\,. det ( A ) ϵ ij = ϵ k l a ki a l j = det ( A T ) ϵ ij = ϵ k l a ik a j l . Now we can compute:
det ( A B ) = ϵ i j ( A B ) 1 i ( A B ) 2 j = ϵ i j ( a 1 k b k i ) ( a 2 l b l j ) = a 1 k a 2 l ( ϵ i j b k i b l j ) = \det(\mathbf{AB})
= \epsilon^{ij} (\mathbf{AB})_{1i} (\mathbf{AB})_{2j}
= \epsilon^{ij} (a_1{}^{k}b_{ki}) (a_2{}^{l}b_{lj})
= a_1{}^{k} a_2{}^{l} (\epsilon^{ij} b_{ki} b_{lj}) = det ( AB ) = ϵ ij ( AB ) 1 i ( AB ) 2 j = ϵ ij ( a 1 k b ki ) ( a 2 l b l j ) = a 1 k a 2 l ( ϵ ij b ki b l j ) =
= a 1 k a 2 l ( ϵ i j b i k b j l ) = a 1 k a 2 l det ( B ) ϵ k l = ( ϵ k l a 1 k a 2 l ) det ( B ) = det ( A ) det ( B ) . = a_1{}^{k} a_2{}^{l} (\epsilon^{ij} b_{ik} b_{jl})
= a_1{}^{k} a_2{}^{l} \det(\mathbf{B}) \epsilon_{kl}
= (\epsilon^{kl} a_{1k} a_{2l}) \det(\mathbf{B})
= \det(\mathbf{A}) \det(\mathbf{B})\,. = a 1 k a 2 l ( ϵ ij b ik b j l ) = a 1 k a 2 l det ( B ) ϵ k l = ( ϵ k l a 1 k a 2 l ) det ( B ) = det ( A ) det ( B ) .