Coefficients of the characteristic polynomialIs Sum of Principal Minors Equals to Pseudo Determinant?Is there any easiest way to find the determinant?How to tackle this polynomial given as a determinant?What is the characteristic polynomial of this matrix?Characteristic polynomial problemCalculate a determinant with patternCalculating general $ntimes n$ determinantHow can we prove the equality of the following determinants?The recursive relation of polynomial characteristic of a matrixProving two matrices are similar using the characteristic polynomialDiagonal entries are zero, others are $1$. Find the determinant.

Multi tool use
Multi tool use

Does ultrasonic bath cleaning damage laboratory volumetric glassware calibration?

Polish letters in ASME English template

Why Is Abelian Gauge Theory So Special?

What is the olden name for sideburns?

Do we or do we not observe (measure) superpositions all the time?

Should I tell my insurance company I have an unsecured loan for my new car?

How exactly is a normal force exerted, at the molecular level?

Is there a short way to check uniqueness of values without using 'if' and multiple 'and's?

Why won't the ground take my seed?

Can I travel from Germany to England alone as an unaccompanied minor?

Do 3D printers really reach 50 micron (0.050mm) accuracy?

Does a centaur PC also count as being mounted?

Does anycast addressing add additional latency in any way?

where clause to retrieve case record which are older than 12 months and 1 day in soql

How do I spend money in Sweden and Denmark?

Do sudoku answers always have a single minimal clue set?

How to convert object fill in to fine lines?

Why does the numerical solution of an ODE move away from an unstable equilibrium?

Why does this function call behave sensibly after calling it through a typecasted function pointer?

Zombie diet, why humans?

When to apply Lorentz transformations and laws of time dilations and length contractions: explanations

Why do I have to press the shutter button twice on my Canon 6d Mark II?

Children's short story about material that accelerates away from gravity

Anagram Within an Anagram!



Coefficients of the characteristic polynomial


Is Sum of Principal Minors Equals to Pseudo Determinant?Is there any easiest way to find the determinant?How to tackle this polynomial given as a determinant?What is the characteristic polynomial of this matrix?Characteristic polynomial problemCalculate a determinant with patternCalculating general $ntimes n$ determinantHow can we prove the equality of the following determinants?The recursive relation of polynomial characteristic of a matrixProving two matrices are similar using the characteristic polynomialDiagonal entries are zero, others are $1$. Find the determinant.






.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty margin-bottom:0;








5












$begingroup$


Let $A$ be an $ntimes n$ matrix. Then its characteristic polynomial is



$$phi_A(x) = det(xI - A) = x^n - (operatornametr A)x^n-1 + cdots + (-1)^n-1(operatornametr(operatornameadjA))x + (-1)^ndet A,$$



where $operatornameadjA$ denotes the adjugate matrix of $A$.




My question: Why are the coefficients of $phi_A$ those shown above?




I can see why $(-1)^ndet A$ is the constant term, since
$$det(A) = phi_A(0) = (0-mu_1)cdots(0-mu_n) = (-1)^nmu_1cdotsmu_n,$$



where $mu_i$ are the roots ($i=1,dots,n$). I also see why the coefficient of $x^n-1$ is minus the trace, since by Laplace expansion (along the first row):



beginalign*
&det(xI-A)\[10pt]
&= beginvmatrix
x-a_11 & -a_12 & cdots & -a_1n\
-a_21 & x-a_22 & cdots & -a_2n\
vdots & vdots & ddots & vdots\
-a_n1 & -a_n2 & cdots & x-a_nn
endvmatrix\[10pt]
&= (x-a_11)beginvmatrix
x-a_22 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix + underbracea_12 beginvmatrix
-a_21 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n1 & cdots & x-a_nn
endvmatrix + cdots_textat most polynomial of degree $n-2$\[10pt]
&=(x-a_11)bigg((x-a_22)beginvmatrix
x-a_33 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n3 & cdots & x-a_nn
endvmatrix
+underbraceunderbracea_23beginvmatrix
-a_32 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix+cdots_textat most polynomial of degree $n-3$
bigg) + cdots_textat most polynomial of degree $n-2$\
&kern10ptvdots\[10pt]
&=(x-a_11)(x-a_22)cdots(x-a_nn) + underbracecdots_textat most polynomial of degree $n-2$\[10pt]
&= x^n - (a_11+a_22+cdots + a_nn)x^n-1 + cdots\[10pt]
&= x^n - (operatornametrA)x^n-1 + cdots
endalign*

(Let me know if there is a more elegant way to obtain this).



But what I cannot seem to prove is that the coefficient of $x$ is $ (-1)^n-1operatornametr(operatornameadjA)$. Attempting to work with the Laplace expansion (as I did to obtain the trace above) is leading me nowhere.










share|cite|improve this question











$endgroup$











  • $begingroup$
    Leibniz’ expansion seems like it was almost made for proving this.
    $endgroup$
    – amd
    6 hours ago

















5












$begingroup$


Let $A$ be an $ntimes n$ matrix. Then its characteristic polynomial is



$$phi_A(x) = det(xI - A) = x^n - (operatornametr A)x^n-1 + cdots + (-1)^n-1(operatornametr(operatornameadjA))x + (-1)^ndet A,$$



where $operatornameadjA$ denotes the adjugate matrix of $A$.




My question: Why are the coefficients of $phi_A$ those shown above?




I can see why $(-1)^ndet A$ is the constant term, since
$$det(A) = phi_A(0) = (0-mu_1)cdots(0-mu_n) = (-1)^nmu_1cdotsmu_n,$$



where $mu_i$ are the roots ($i=1,dots,n$). I also see why the coefficient of $x^n-1$ is minus the trace, since by Laplace expansion (along the first row):



beginalign*
&det(xI-A)\[10pt]
&= beginvmatrix
x-a_11 & -a_12 & cdots & -a_1n\
-a_21 & x-a_22 & cdots & -a_2n\
vdots & vdots & ddots & vdots\
-a_n1 & -a_n2 & cdots & x-a_nn
endvmatrix\[10pt]
&= (x-a_11)beginvmatrix
x-a_22 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix + underbracea_12 beginvmatrix
-a_21 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n1 & cdots & x-a_nn
endvmatrix + cdots_textat most polynomial of degree $n-2$\[10pt]
&=(x-a_11)bigg((x-a_22)beginvmatrix
x-a_33 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n3 & cdots & x-a_nn
endvmatrix
+underbraceunderbracea_23beginvmatrix
-a_32 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix+cdots_textat most polynomial of degree $n-3$
bigg) + cdots_textat most polynomial of degree $n-2$\
&kern10ptvdots\[10pt]
&=(x-a_11)(x-a_22)cdots(x-a_nn) + underbracecdots_textat most polynomial of degree $n-2$\[10pt]
&= x^n - (a_11+a_22+cdots + a_nn)x^n-1 + cdots\[10pt]
&= x^n - (operatornametrA)x^n-1 + cdots
endalign*

(Let me know if there is a more elegant way to obtain this).



But what I cannot seem to prove is that the coefficient of $x$ is $ (-1)^n-1operatornametr(operatornameadjA)$. Attempting to work with the Laplace expansion (as I did to obtain the trace above) is leading me nowhere.










share|cite|improve this question











$endgroup$











  • $begingroup$
    Leibniz’ expansion seems like it was almost made for proving this.
    $endgroup$
    – amd
    6 hours ago













5












5








5


2



$begingroup$


Let $A$ be an $ntimes n$ matrix. Then its characteristic polynomial is



$$phi_A(x) = det(xI - A) = x^n - (operatornametr A)x^n-1 + cdots + (-1)^n-1(operatornametr(operatornameadjA))x + (-1)^ndet A,$$



where $operatornameadjA$ denotes the adjugate matrix of $A$.




My question: Why are the coefficients of $phi_A$ those shown above?




I can see why $(-1)^ndet A$ is the constant term, since
$$det(A) = phi_A(0) = (0-mu_1)cdots(0-mu_n) = (-1)^nmu_1cdotsmu_n,$$



where $mu_i$ are the roots ($i=1,dots,n$). I also see why the coefficient of $x^n-1$ is minus the trace, since by Laplace expansion (along the first row):



beginalign*
&det(xI-A)\[10pt]
&= beginvmatrix
x-a_11 & -a_12 & cdots & -a_1n\
-a_21 & x-a_22 & cdots & -a_2n\
vdots & vdots & ddots & vdots\
-a_n1 & -a_n2 & cdots & x-a_nn
endvmatrix\[10pt]
&= (x-a_11)beginvmatrix
x-a_22 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix + underbracea_12 beginvmatrix
-a_21 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n1 & cdots & x-a_nn
endvmatrix + cdots_textat most polynomial of degree $n-2$\[10pt]
&=(x-a_11)bigg((x-a_22)beginvmatrix
x-a_33 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n3 & cdots & x-a_nn
endvmatrix
+underbraceunderbracea_23beginvmatrix
-a_32 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix+cdots_textat most polynomial of degree $n-3$
bigg) + cdots_textat most polynomial of degree $n-2$\
&kern10ptvdots\[10pt]
&=(x-a_11)(x-a_22)cdots(x-a_nn) + underbracecdots_textat most polynomial of degree $n-2$\[10pt]
&= x^n - (a_11+a_22+cdots + a_nn)x^n-1 + cdots\[10pt]
&= x^n - (operatornametrA)x^n-1 + cdots
endalign*

(Let me know if there is a more elegant way to obtain this).



But what I cannot seem to prove is that the coefficient of $x$ is $ (-1)^n-1operatornametr(operatornameadjA)$. Attempting to work with the Laplace expansion (as I did to obtain the trace above) is leading me nowhere.










share|cite|improve this question











$endgroup$




Let $A$ be an $ntimes n$ matrix. Then its characteristic polynomial is



$$phi_A(x) = det(xI - A) = x^n - (operatornametr A)x^n-1 + cdots + (-1)^n-1(operatornametr(operatornameadjA))x + (-1)^ndet A,$$



where $operatornameadjA$ denotes the adjugate matrix of $A$.




My question: Why are the coefficients of $phi_A$ those shown above?




I can see why $(-1)^ndet A$ is the constant term, since
$$det(A) = phi_A(0) = (0-mu_1)cdots(0-mu_n) = (-1)^nmu_1cdotsmu_n,$$



where $mu_i$ are the roots ($i=1,dots,n$). I also see why the coefficient of $x^n-1$ is minus the trace, since by Laplace expansion (along the first row):



beginalign*
&det(xI-A)\[10pt]
&= beginvmatrix
x-a_11 & -a_12 & cdots & -a_1n\
-a_21 & x-a_22 & cdots & -a_2n\
vdots & vdots & ddots & vdots\
-a_n1 & -a_n2 & cdots & x-a_nn
endvmatrix\[10pt]
&= (x-a_11)beginvmatrix
x-a_22 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix + underbracea_12 beginvmatrix
-a_21 & cdots & -a_2n\
vdots & ddots & vdots\
-a_n1 & cdots & x-a_nn
endvmatrix + cdots_textat most polynomial of degree $n-2$\[10pt]
&=(x-a_11)bigg((x-a_22)beginvmatrix
x-a_33 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n3 & cdots & x-a_nn
endvmatrix
+underbraceunderbracea_23beginvmatrix
-a_32 & cdots & -a_3n\
vdots & ddots & vdots\
-a_n2 & cdots & x-a_nn
endvmatrix+cdots_textat most polynomial of degree $n-3$
bigg) + cdots_textat most polynomial of degree $n-2$\
&kern10ptvdots\[10pt]
&=(x-a_11)(x-a_22)cdots(x-a_nn) + underbracecdots_textat most polynomial of degree $n-2$\[10pt]
&= x^n - (a_11+a_22+cdots + a_nn)x^n-1 + cdots\[10pt]
&= x^n - (operatornametrA)x^n-1 + cdots
endalign*

(Let me know if there is a more elegant way to obtain this).



But what I cannot seem to prove is that the coefficient of $x$ is $ (-1)^n-1operatornametr(operatornameadjA)$. Attempting to work with the Laplace expansion (as I did to obtain the trace above) is leading me nowhere.







linear-algebra matrices determinant






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited 8 hours ago







Luke Collins

















asked 8 hours ago









Luke CollinsLuke Collins

1,5364 silver badges21 bronze badges




1,5364 silver badges21 bronze badges











  • $begingroup$
    Leibniz’ expansion seems like it was almost made for proving this.
    $endgroup$
    – amd
    6 hours ago
















  • $begingroup$
    Leibniz’ expansion seems like it was almost made for proving this.
    $endgroup$
    – amd
    6 hours ago















$begingroup$
Leibniz’ expansion seems like it was almost made for proving this.
$endgroup$
– amd
6 hours ago




$begingroup$
Leibniz’ expansion seems like it was almost made for proving this.
$endgroup$
– amd
6 hours ago










3 Answers
3






active

oldest

votes


















2












$begingroup$

This can easily be derived from Corollary 2 in https://math.stackexchange.com/a/3069061/ (applied to $-A$ instead of $A$) by setting $r = 1$, $r = n-1$ or $r = n$. If you follow the reference to my notes, you'll see that I basically make the "expand everything and extract the coefficients of the appropriate powers of $x$" argument that you already did for the $x^n-2$-coefficient and that @J_P sketched for the other coefficients; all I am doing differently is actually formalizing it.



For another proof, see my note The trace Cayley-Hamilton theorem (temporary mirror while the website just linked is offline), specifically:



  • Corollary 2.4 (b) for the $x^0$-coefficient being $left(-1right)^n det A$;


  • Theorem 5.32 for the $x^1$-coefficient being $left(-1right)^n-1 operatornameTrleft(operatornameadjAright)$;


  • Corollary 3.22 for the $x^n-1$-coefficient being $- operatornameTr A$.


Note that my goal in that note (at least the original goal; feature creep has borne all of Section 5 and parts of the preceding sections) is not to find the coefficients of the characteristic polynomial, so I am not taking any sort of beeline. There isn't much to say in favor of my proof other than that it avoids the handwaving that typically appears in the standard (term-by-term expansion) argument.






share|cite|improve this answer









$endgroup$












  • $begingroup$
    I've always wanted to see this done more rigorously, I'll have to sit down and peruse your other post carefully. Apparently this can also be done with exterior powers but I don't know the first thing about them, do you have any references for that subject?
    $endgroup$
    – J_P
    7 hours ago











  • $begingroup$
    I also tried my own hand at formalizing the argument which I added to my answer. I'd be very grateful if you found the time to just quickly scan it and tell me if it seems okay.
    $endgroup$
    – J_P
    6 hours ago










  • $begingroup$
    @J_P: Ironically, it is the "This can be rigorously demonstrated by observing" part that I don't understand about your proof, but it is otherwise a nice outline of a rigorous proof without too much pain (and the part I don't understand can be done by induction). Good job!
    $endgroup$
    – darij grinberg
    1 hour ago


















1












$begingroup$

Suppose first $B$ is an $n times n$ matrix with $det(B) = 1$. Then $textadj(B) = B^-1$, and



$$ phi_B^-1(x) = det(x I - B^-1) = det(x B - I) = (-x)^n det(x^-1 I - B)
= (-x)^n phi_B(1/x)$$



The coefficient of $x^1$ in $phi_B(x)$ is the coefficient of $x^-1$ in $phi_B(1/x)$, i.e. $(-1)^n$ times the coefficient of $x^n-1$ in $(-x)^n phi_B(1/x)$, but you know the coefficient of $x^n-1$ in $phi_B^-1(x)$ is
$texttr(B^-1)$. That is, the coefficient of $x^1$ in $phi_B(x)$ is
$(-1)^n texttr(B^-1) = (-1)^n texttr(textadj; B)$.



Now let $A = t B$, where $t ne 0$. Then $phi_A(x) = det(x I - t B) = t^n det((x/t)I - A) = t^n phi_B(x/t)$. The coefficient of $x^1$ in $phi_A(x)$ is then
$t^n-1$ times the coefficient of $x^1$ in $phi_B(x)$. But also
$textadj; A = t^n-1 textadj; B$. So we again obtain that the coefficient of
$x^1$ in $phi_A(x)$ is $(-1)^n texttr(textadj; A)$.



Every nonsingular matrix
$A = det(A)^1/n B$ where $det(B) = 1$, so the formula for the coefficient holds for every nonsingular matrix. (Note that in the case where $det(A)$ is negative and $n$ is even, we must use complex numbers here, but that's not a problem)



Now the coefficient of $x^1$ in $phi_A(x)$ and $(-1)^n texttr(textadj; A)$ are both polynomials in the coefficients of $A$.

Thus the equation must hold for all $n times n$ matrices, not just the nonsingular ones.






share|cite|improve this answer









$endgroup$




















    0












    $begingroup$

    I define $phi_A(x)=det(A-lambda I)$. You get the coefficient at $lambda^n-k$ in the following way: pick $k$ diagonal entries $a_i_1i_1,...,a_i_ki_k$. These define a $ktimes k$ minor of $A$ that you get by removing every row and column that does not contain one of the chosen diagonal elements. Calculate the determinant of this minor and sum over all $nchoose k$ selections of the $k$ diagonal entries. Up to sign (which is just alternating) that is your $lambda^n-k$ coefficient.



    I have to say that I don't know a precise formal proof (EDIT: go to the bottom of the answer, I've now added an attempt at formalizing what follows), I like to view this in a more algorithmic fashion. Here's how I convince myself:
    $$
    beginvmatrix
    a_11-lambda & a_12 & ... & a_1n \
    a_21 & a_22-lambda & ... & a_2n \
    ... & ... & ... & ... \
    a_n1 & a_n2 & ...& a_nn-lambda
    endvmatrix=\
    =beginvmatrix
    a_11 & a_12 & ... & a_1n \
    a_21 & a_22-lambda & ... & a_2n \
    ... & ... & ... & ... \
    a_n1 & a_n2 & ...& a_nn-lambda
    endvmatrix-lambdabeginvmatrix
    a_22-lambda & ... & a_2n \
    ... & ... & ... \
    a_n2 & ...& a_nn-lambda
    endvmatrix
    $$

    Now you see the problem has split into two parts. Deal with each one seperately: in the first one, move on to the next column and again split the determinant up as
    $$
    beginvmatrix
    a_11 & a_12 & ... & a_1n \
    a_21 & a_22-lambda & ... & a_2n \
    ... & ... & ... & ... \
    a_n1 & a_n2 & ...& a_nn-lambda
    endvmatrix=beginvmatrix
    a_11 & a_12 & ... & a_1n \
    a_21 & a_22 & ... & a_2n \
    ... & ... & ... & ... \
    a_n1 & a_n2 & ...& a_nn-lambda
    endvmatrix-lambdabeginvmatrix
    a_11 & ... & a_2n \
    ... & ... & ... \
    a_n2 & ...& a_nn-lambda
    endvmatrix
    $$

    Do the same for the second one. As you keep doing this, a huge number of determinants will crop up ($2^n$ in total - which is just the sum of the binomial coefficients $nchoose k$, already suggestive). You can view this as a decision tree: at each step, you say whether you want to keep the $i$-th column or not. After you pick, you won't touch the previous columns again. So doing this will give you square determinants with certain columns missing. If you pick some columns (equivalently, some diagonal elements) to remove, say "NO" to every decision where you remove one of the desired columns and "YES" otherwise. This defines a path down the binary tree which terminantes when no more $lambda$'s are left. Since every selection of $k$ diagonal elements corresponds to exactly one path down the binary tree with $n-k$ "YES"'s, you will get $(-1)^n-klambda^n-k$ in front of each such determinant and so when you sum it all up, you get precisely the sum of of all diagonal-centered $ktimes k$ determinants.



    In particular, you get the linear term by removing only $1$ column and row and summing over all possibilities. Since these are determinants of the principal minors which feature in the definition of the adjugate matrix, what you calculate is just the sum of the diagonal elements of the adjugate - that is, its trace.



    If you don't believe all this mumbo-jumbo (which is perfectly reasonable), I strongly encourage you to try a $4times 4$ case since that is only $16$ determinants in total, or, if you have a lot of paper, the $5times 5$ case. That might help you get a "feel" for the pattern.



    I also think this exact argument could be made more rigorous if you threw in some arguments like "at the $i$-th step, determinants of a certain size will appear at a certain power of $lambda$" and try doing this by induction, but I never actually went through with this so it's just a hunch at this point.




    So I tried my hand at formalizing this a bit and here's my attempt. Let $D(b_1,b_2,...,b_n)$ denote the various determinants that will appear, where $b_iin0,1,2$. $b_i=0$ means that you remove the $i$-th row and column entirely, $b_i=1$ means that you only remove the $lambda$ at $a_ii$ and $b_i=2$ means "leave the $i$-th diagonal element and its row and column as is". For example:
    $$
    D(2,2,2)=beginvmatrix
    a_11-lambda & a_12 & a_13 \
    a_21 & a_22-lambda & a_23 \
    a_31 & a_32 & a_33-lambda
    endvmatrixquad
    D(2,1,1)=beginvmatrix
    a_11-lambda & a_12 & a_13 \
    a_21 & a_22 & a_23 \
    a_31 & a_32 & a_33
    endvmatrix\
    D(2,1,0)=beginvmatrix
    a_11-lambda & a_12\
    a_21 & a_22
    endvmatrix
    $$

    We have the reduction formula: $$D(b_1,...,2,...,b_n)=D(b_1,...,1,...,b_n)-lambda D(b_1,...,0,...,b_n)$$
    We start with $D(2_1,2_2,...,2_n)$. Write $bfb$ for the vector of $b_i$ and let $Vert bVert$ be the number of $1$'s in $bfb$. If we proceed by induction, we can say that
    $$D(0,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(0,bf b)(-lambda)^n-1-Vertbf bVert$$
    Since the $0$ at the front plays no role and the reduction formula is the same regardless of what the first $b_i$ is, we must have that also:
    $$D(1,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(1,bf b)(-lambda)^n-1-Vertbf bVert$$
    This can be rigorously demonstrated by observing that if we apply the reduction formula to $D(0,bf b)$ on some index, we get the same as if we would for $D(1text or 2,bf b)$ but for the first index. By applying reduction inductively on the indices, we eliminate all $2$'s but since at each step the formula is the same except that we change the first index, after all's said and done, we're left with identical formulas but for the first index.

    Anyway, we have:
    $$
    D(2_1,2_2,...,2_n)=D(1,2,...,2)-lambda D(0,2,...,2)=\
    =sum_dimbf b=n-1\quad ,2notinbf bleft(D(1,bf b)(-lambda)^n-1-Vertbf bVert+D(0,bf b)(-lambda)^n-Vertbf bVertright)=sum_dimbf b=n\,,,,, 2notinbf bD(bf b)(-lambda)^n-Vertbf bVert
    $$

    This establishes the inductive step and completes the proof.




    Due to the helpful suggestion of @amd in the comments, here's an approach with Leibniz's formula. Here, $N=1,...,n$.
    beginalign
    &det(A-lambda I)=sum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_i=1^n(a_isigma(i)-lambdadelta_isigma(i))=\
    &=sum_sigmainmathrmSym(n)mathrmsgn(sigma)sum_Asubset Nprod_iin Aa_isigma(i)prod_jin N-A(-lambda)delta_jsigma(j)=\
    &=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_iin Aa_isigma(i)prod_jin N-Adelta_jsigma(j)
    endalign

    In the last expression we have that $prod_jin N-Adelta_jsigma(j)$ which is $0$ unless $N-A$ contains only fixed points of $sigma$ - in other words, we need only sum over $sigma$ whose fixed points are a superset of $N-A$. But these new permutations form a group which is essentially just $mathrmSym(A)$, the permutations of $A$. The sign of a permutation of length $m$ is $(-1)^k$ where $k=m-text(number of disjoint cycles that make up $sigma$)$ - when we drop the stationary points from $sigma$, $m$ decreases but the number of disjoint cycles decreases by the exact same amount since stationary points correspond to trivial cycles. Hence the sign remains the same and we have:
    $$
    det(A-lambda I)=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(A)mathrmsgn(sigma)prod_iin Aa_isigma(i)
    $$

    But the sums over $sigmainmathrmSym(A)$ are now just determinants of the principal minors and we are done.






    share|cite|improve this answer











    $endgroup$












    • $begingroup$
      I believe that the Leibniz formula for determinants can be used to prove this.
      $endgroup$
      – amd
      7 hours ago










    • $begingroup$
      Possibly, but I dread the thought of using that formula, I just find it really unwieldy.
      $endgroup$
      – J_P
      6 hours ago











    • $begingroup$
      It’s awful for computations, but great for proofs. In particular, your construction is equivalent to picking out the terms that involve exactly $n-k$ $lambda$’s.
      $endgroup$
      – amd
      6 hours ago










    • $begingroup$
      Wow, you're right! I think I managed to do it. Well, I'm not absolutely certain about my combinatorics (have to think about permutations a bit at one step) but I think it works. I'll also add this approach to my answer unless you were thinking of posting it yourself in which case tell me and I'll delete my edit.
      $endgroup$
      – J_P
      5 hours ago










    • $begingroup$
      You need to be a bit careful about contributions from terms with more $lambda$’s, but they might cancel nicely. Anyway, that formula is where I’d start looking for a proof.
      $endgroup$
      – amd
      5 hours ago














    Your Answer








    StackExchange.ready(function()
    var channelOptions =
    tags: "".split(" "),
    id: "69"
    ;
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function()
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled)
    StackExchange.using("snippets", function()
    createEditor();
    );

    else
    createEditor();

    );

    function createEditor()
    StackExchange.prepareEditor(
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader:
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    ,
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    );



    );













    draft saved

    draft discarded


















    StackExchange.ready(
    function ()
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3272063%2fcoefficients-of-the-characteristic-polynomial%23new-answer', 'question_page');

    );

    Post as a guest















    Required, but never shown

























    3 Answers
    3






    active

    oldest

    votes








    3 Answers
    3






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    2












    $begingroup$

    This can easily be derived from Corollary 2 in https://math.stackexchange.com/a/3069061/ (applied to $-A$ instead of $A$) by setting $r = 1$, $r = n-1$ or $r = n$. If you follow the reference to my notes, you'll see that I basically make the "expand everything and extract the coefficients of the appropriate powers of $x$" argument that you already did for the $x^n-2$-coefficient and that @J_P sketched for the other coefficients; all I am doing differently is actually formalizing it.



    For another proof, see my note The trace Cayley-Hamilton theorem (temporary mirror while the website just linked is offline), specifically:



    • Corollary 2.4 (b) for the $x^0$-coefficient being $left(-1right)^n det A$;


    • Theorem 5.32 for the $x^1$-coefficient being $left(-1right)^n-1 operatornameTrleft(operatornameadjAright)$;


    • Corollary 3.22 for the $x^n-1$-coefficient being $- operatornameTr A$.


    Note that my goal in that note (at least the original goal; feature creep has borne all of Section 5 and parts of the preceding sections) is not to find the coefficients of the characteristic polynomial, so I am not taking any sort of beeline. There isn't much to say in favor of my proof other than that it avoids the handwaving that typically appears in the standard (term-by-term expansion) argument.






    share|cite|improve this answer









    $endgroup$












    • $begingroup$
      I've always wanted to see this done more rigorously, I'll have to sit down and peruse your other post carefully. Apparently this can also be done with exterior powers but I don't know the first thing about them, do you have any references for that subject?
      $endgroup$
      – J_P
      7 hours ago











    • $begingroup$
      I also tried my own hand at formalizing the argument which I added to my answer. I'd be very grateful if you found the time to just quickly scan it and tell me if it seems okay.
      $endgroup$
      – J_P
      6 hours ago










    • $begingroup$
      @J_P: Ironically, it is the "This can be rigorously demonstrated by observing" part that I don't understand about your proof, but it is otherwise a nice outline of a rigorous proof without too much pain (and the part I don't understand can be done by induction). Good job!
      $endgroup$
      – darij grinberg
      1 hour ago















    2












    $begingroup$

    This can easily be derived from Corollary 2 in https://math.stackexchange.com/a/3069061/ (applied to $-A$ instead of $A$) by setting $r = 1$, $r = n-1$ or $r = n$. If you follow the reference to my notes, you'll see that I basically make the "expand everything and extract the coefficients of the appropriate powers of $x$" argument that you already did for the $x^n-2$-coefficient and that @J_P sketched for the other coefficients; all I am doing differently is actually formalizing it.



    For another proof, see my note The trace Cayley-Hamilton theorem (temporary mirror while the website just linked is offline), specifically:



    • Corollary 2.4 (b) for the $x^0$-coefficient being $left(-1right)^n det A$;


    • Theorem 5.32 for the $x^1$-coefficient being $left(-1right)^n-1 operatornameTrleft(operatornameadjAright)$;


    • Corollary 3.22 for the $x^n-1$-coefficient being $- operatornameTr A$.


    Note that my goal in that note (at least the original goal; feature creep has borne all of Section 5 and parts of the preceding sections) is not to find the coefficients of the characteristic polynomial, so I am not taking any sort of beeline. There isn't much to say in favor of my proof other than that it avoids the handwaving that typically appears in the standard (term-by-term expansion) argument.






    share|cite|improve this answer









    $endgroup$












    • $begingroup$
      I've always wanted to see this done more rigorously, I'll have to sit down and peruse your other post carefully. Apparently this can also be done with exterior powers but I don't know the first thing about them, do you have any references for that subject?
      $endgroup$
      – J_P
      7 hours ago











    • $begingroup$
      I also tried my own hand at formalizing the argument which I added to my answer. I'd be very grateful if you found the time to just quickly scan it and tell me if it seems okay.
      $endgroup$
      – J_P
      6 hours ago










    • $begingroup$
      @J_P: Ironically, it is the "This can be rigorously demonstrated by observing" part that I don't understand about your proof, but it is otherwise a nice outline of a rigorous proof without too much pain (and the part I don't understand can be done by induction). Good job!
      $endgroup$
      – darij grinberg
      1 hour ago













    2












    2








    2





    $begingroup$

    This can easily be derived from Corollary 2 in https://math.stackexchange.com/a/3069061/ (applied to $-A$ instead of $A$) by setting $r = 1$, $r = n-1$ or $r = n$. If you follow the reference to my notes, you'll see that I basically make the "expand everything and extract the coefficients of the appropriate powers of $x$" argument that you already did for the $x^n-2$-coefficient and that @J_P sketched for the other coefficients; all I am doing differently is actually formalizing it.



    For another proof, see my note The trace Cayley-Hamilton theorem (temporary mirror while the website just linked is offline), specifically:



    • Corollary 2.4 (b) for the $x^0$-coefficient being $left(-1right)^n det A$;


    • Theorem 5.32 for the $x^1$-coefficient being $left(-1right)^n-1 operatornameTrleft(operatornameadjAright)$;


    • Corollary 3.22 for the $x^n-1$-coefficient being $- operatornameTr A$.


    Note that my goal in that note (at least the original goal; feature creep has borne all of Section 5 and parts of the preceding sections) is not to find the coefficients of the characteristic polynomial, so I am not taking any sort of beeline. There isn't much to say in favor of my proof other than that it avoids the handwaving that typically appears in the standard (term-by-term expansion) argument.






    share|cite|improve this answer









    $endgroup$



    This can easily be derived from Corollary 2 in https://math.stackexchange.com/a/3069061/ (applied to $-A$ instead of $A$) by setting $r = 1$, $r = n-1$ or $r = n$. If you follow the reference to my notes, you'll see that I basically make the "expand everything and extract the coefficients of the appropriate powers of $x$" argument that you already did for the $x^n-2$-coefficient and that @J_P sketched for the other coefficients; all I am doing differently is actually formalizing it.



    For another proof, see my note The trace Cayley-Hamilton theorem (temporary mirror while the website just linked is offline), specifically:



    • Corollary 2.4 (b) for the $x^0$-coefficient being $left(-1right)^n det A$;


    • Theorem 5.32 for the $x^1$-coefficient being $left(-1right)^n-1 operatornameTrleft(operatornameadjAright)$;


    • Corollary 3.22 for the $x^n-1$-coefficient being $- operatornameTr A$.


    Note that my goal in that note (at least the original goal; feature creep has borne all of Section 5 and parts of the preceding sections) is not to find the coefficients of the characteristic polynomial, so I am not taking any sort of beeline. There isn't much to say in favor of my proof other than that it avoids the handwaving that typically appears in the standard (term-by-term expansion) argument.







    share|cite|improve this answer












    share|cite|improve this answer



    share|cite|improve this answer










    answered 7 hours ago









    darij grinbergdarij grinberg

    12k4 gold badges32 silver badges69 bronze badges




    12k4 gold badges32 silver badges69 bronze badges











    • $begingroup$
      I've always wanted to see this done more rigorously, I'll have to sit down and peruse your other post carefully. Apparently this can also be done with exterior powers but I don't know the first thing about them, do you have any references for that subject?
      $endgroup$
      – J_P
      7 hours ago











    • $begingroup$
      I also tried my own hand at formalizing the argument which I added to my answer. I'd be very grateful if you found the time to just quickly scan it and tell me if it seems okay.
      $endgroup$
      – J_P
      6 hours ago










    • $begingroup$
      @J_P: Ironically, it is the "This can be rigorously demonstrated by observing" part that I don't understand about your proof, but it is otherwise a nice outline of a rigorous proof without too much pain (and the part I don't understand can be done by induction). Good job!
      $endgroup$
      – darij grinberg
      1 hour ago
















    • $begingroup$
      I've always wanted to see this done more rigorously, I'll have to sit down and peruse your other post carefully. Apparently this can also be done with exterior powers but I don't know the first thing about them, do you have any references for that subject?
      $endgroup$
      – J_P
      7 hours ago











    • $begingroup$
      I also tried my own hand at formalizing the argument which I added to my answer. I'd be very grateful if you found the time to just quickly scan it and tell me if it seems okay.
      $endgroup$
      – J_P
      6 hours ago










    • $begingroup$
      @J_P: Ironically, it is the "This can be rigorously demonstrated by observing" part that I don't understand about your proof, but it is otherwise a nice outline of a rigorous proof without too much pain (and the part I don't understand can be done by induction). Good job!
      $endgroup$
      – darij grinberg
      1 hour ago















    $begingroup$
    I've always wanted to see this done more rigorously, I'll have to sit down and peruse your other post carefully. Apparently this can also be done with exterior powers but I don't know the first thing about them, do you have any references for that subject?
    $endgroup$
    – J_P
    7 hours ago





    $begingroup$
    I've always wanted to see this done more rigorously, I'll have to sit down and peruse your other post carefully. Apparently this can also be done with exterior powers but I don't know the first thing about them, do you have any references for that subject?
    $endgroup$
    – J_P
    7 hours ago













    $begingroup$
    I also tried my own hand at formalizing the argument which I added to my answer. I'd be very grateful if you found the time to just quickly scan it and tell me if it seems okay.
    $endgroup$
    – J_P
    6 hours ago




    $begingroup$
    I also tried my own hand at formalizing the argument which I added to my answer. I'd be very grateful if you found the time to just quickly scan it and tell me if it seems okay.
    $endgroup$
    – J_P
    6 hours ago












    $begingroup$
    @J_P: Ironically, it is the "This can be rigorously demonstrated by observing" part that I don't understand about your proof, but it is otherwise a nice outline of a rigorous proof without too much pain (and the part I don't understand can be done by induction). Good job!
    $endgroup$
    – darij grinberg
    1 hour ago




    $begingroup$
    @J_P: Ironically, it is the "This can be rigorously demonstrated by observing" part that I don't understand about your proof, but it is otherwise a nice outline of a rigorous proof without too much pain (and the part I don't understand can be done by induction). Good job!
    $endgroup$
    – darij grinberg
    1 hour ago













    1












    $begingroup$

    Suppose first $B$ is an $n times n$ matrix with $det(B) = 1$. Then $textadj(B) = B^-1$, and



    $$ phi_B^-1(x) = det(x I - B^-1) = det(x B - I) = (-x)^n det(x^-1 I - B)
    = (-x)^n phi_B(1/x)$$



    The coefficient of $x^1$ in $phi_B(x)$ is the coefficient of $x^-1$ in $phi_B(1/x)$, i.e. $(-1)^n$ times the coefficient of $x^n-1$ in $(-x)^n phi_B(1/x)$, but you know the coefficient of $x^n-1$ in $phi_B^-1(x)$ is
    $texttr(B^-1)$. That is, the coefficient of $x^1$ in $phi_B(x)$ is
    $(-1)^n texttr(B^-1) = (-1)^n texttr(textadj; B)$.



    Now let $A = t B$, where $t ne 0$. Then $phi_A(x) = det(x I - t B) = t^n det((x/t)I - A) = t^n phi_B(x/t)$. The coefficient of $x^1$ in $phi_A(x)$ is then
    $t^n-1$ times the coefficient of $x^1$ in $phi_B(x)$. But also
    $textadj; A = t^n-1 textadj; B$. So we again obtain that the coefficient of
    $x^1$ in $phi_A(x)$ is $(-1)^n texttr(textadj; A)$.



    Every nonsingular matrix
    $A = det(A)^1/n B$ where $det(B) = 1$, so the formula for the coefficient holds for every nonsingular matrix. (Note that in the case where $det(A)$ is negative and $n$ is even, we must use complex numbers here, but that's not a problem)



    Now the coefficient of $x^1$ in $phi_A(x)$ and $(-1)^n texttr(textadj; A)$ are both polynomials in the coefficients of $A$.

    Thus the equation must hold for all $n times n$ matrices, not just the nonsingular ones.






    share|cite|improve this answer









    $endgroup$

















      1












      $begingroup$

      Suppose first $B$ is an $n times n$ matrix with $det(B) = 1$. Then $textadj(B) = B^-1$, and



      $$ phi_B^-1(x) = det(x I - B^-1) = det(x B - I) = (-x)^n det(x^-1 I - B)
      = (-x)^n phi_B(1/x)$$



      The coefficient of $x^1$ in $phi_B(x)$ is the coefficient of $x^-1$ in $phi_B(1/x)$, i.e. $(-1)^n$ times the coefficient of $x^n-1$ in $(-x)^n phi_B(1/x)$, but you know the coefficient of $x^n-1$ in $phi_B^-1(x)$ is
      $texttr(B^-1)$. That is, the coefficient of $x^1$ in $phi_B(x)$ is
      $(-1)^n texttr(B^-1) = (-1)^n texttr(textadj; B)$.



      Now let $A = t B$, where $t ne 0$. Then $phi_A(x) = det(x I - t B) = t^n det((x/t)I - A) = t^n phi_B(x/t)$. The coefficient of $x^1$ in $phi_A(x)$ is then
      $t^n-1$ times the coefficient of $x^1$ in $phi_B(x)$. But also
      $textadj; A = t^n-1 textadj; B$. So we again obtain that the coefficient of
      $x^1$ in $phi_A(x)$ is $(-1)^n texttr(textadj; A)$.



      Every nonsingular matrix
      $A = det(A)^1/n B$ where $det(B) = 1$, so the formula for the coefficient holds for every nonsingular matrix. (Note that in the case where $det(A)$ is negative and $n$ is even, we must use complex numbers here, but that's not a problem)



      Now the coefficient of $x^1$ in $phi_A(x)$ and $(-1)^n texttr(textadj; A)$ are both polynomials in the coefficients of $A$.

      Thus the equation must hold for all $n times n$ matrices, not just the nonsingular ones.






      share|cite|improve this answer









      $endgroup$















        1












        1








        1





        $begingroup$

        Suppose first $B$ is an $n times n$ matrix with $det(B) = 1$. Then $textadj(B) = B^-1$, and



        $$ phi_B^-1(x) = det(x I - B^-1) = det(x B - I) = (-x)^n det(x^-1 I - B)
        = (-x)^n phi_B(1/x)$$



        The coefficient of $x^1$ in $phi_B(x)$ is the coefficient of $x^-1$ in $phi_B(1/x)$, i.e. $(-1)^n$ times the coefficient of $x^n-1$ in $(-x)^n phi_B(1/x)$, but you know the coefficient of $x^n-1$ in $phi_B^-1(x)$ is
        $texttr(B^-1)$. That is, the coefficient of $x^1$ in $phi_B(x)$ is
        $(-1)^n texttr(B^-1) = (-1)^n texttr(textadj; B)$.



        Now let $A = t B$, where $t ne 0$. Then $phi_A(x) = det(x I - t B) = t^n det((x/t)I - A) = t^n phi_B(x/t)$. The coefficient of $x^1$ in $phi_A(x)$ is then
        $t^n-1$ times the coefficient of $x^1$ in $phi_B(x)$. But also
        $textadj; A = t^n-1 textadj; B$. So we again obtain that the coefficient of
        $x^1$ in $phi_A(x)$ is $(-1)^n texttr(textadj; A)$.



        Every nonsingular matrix
        $A = det(A)^1/n B$ where $det(B) = 1$, so the formula for the coefficient holds for every nonsingular matrix. (Note that in the case where $det(A)$ is negative and $n$ is even, we must use complex numbers here, but that's not a problem)



        Now the coefficient of $x^1$ in $phi_A(x)$ and $(-1)^n texttr(textadj; A)$ are both polynomials in the coefficients of $A$.

        Thus the equation must hold for all $n times n$ matrices, not just the nonsingular ones.






        share|cite|improve this answer









        $endgroup$



        Suppose first $B$ is an $n times n$ matrix with $det(B) = 1$. Then $textadj(B) = B^-1$, and



        $$ phi_B^-1(x) = det(x I - B^-1) = det(x B - I) = (-x)^n det(x^-1 I - B)
        = (-x)^n phi_B(1/x)$$



        The coefficient of $x^1$ in $phi_B(x)$ is the coefficient of $x^-1$ in $phi_B(1/x)$, i.e. $(-1)^n$ times the coefficient of $x^n-1$ in $(-x)^n phi_B(1/x)$, but you know the coefficient of $x^n-1$ in $phi_B^-1(x)$ is
        $texttr(B^-1)$. That is, the coefficient of $x^1$ in $phi_B(x)$ is
        $(-1)^n texttr(B^-1) = (-1)^n texttr(textadj; B)$.



        Now let $A = t B$, where $t ne 0$. Then $phi_A(x) = det(x I - t B) = t^n det((x/t)I - A) = t^n phi_B(x/t)$. The coefficient of $x^1$ in $phi_A(x)$ is then
        $t^n-1$ times the coefficient of $x^1$ in $phi_B(x)$. But also
        $textadj; A = t^n-1 textadj; B$. So we again obtain that the coefficient of
        $x^1$ in $phi_A(x)$ is $(-1)^n texttr(textadj; A)$.



        Every nonsingular matrix
        $A = det(A)^1/n B$ where $det(B) = 1$, so the formula for the coefficient holds for every nonsingular matrix. (Note that in the case where $det(A)$ is negative and $n$ is even, we must use complex numbers here, but that's not a problem)



        Now the coefficient of $x^1$ in $phi_A(x)$ and $(-1)^n texttr(textadj; A)$ are both polynomials in the coefficients of $A$.

        Thus the equation must hold for all $n times n$ matrices, not just the nonsingular ones.







        share|cite|improve this answer












        share|cite|improve this answer



        share|cite|improve this answer










        answered 7 hours ago









        Robert IsraelRobert Israel

        340k23 gold badges233 silver badges492 bronze badges




        340k23 gold badges233 silver badges492 bronze badges





















            0












            $begingroup$

            I define $phi_A(x)=det(A-lambda I)$. You get the coefficient at $lambda^n-k$ in the following way: pick $k$ diagonal entries $a_i_1i_1,...,a_i_ki_k$. These define a $ktimes k$ minor of $A$ that you get by removing every row and column that does not contain one of the chosen diagonal elements. Calculate the determinant of this minor and sum over all $nchoose k$ selections of the $k$ diagonal entries. Up to sign (which is just alternating) that is your $lambda^n-k$ coefficient.



            I have to say that I don't know a precise formal proof (EDIT: go to the bottom of the answer, I've now added an attempt at formalizing what follows), I like to view this in a more algorithmic fashion. Here's how I convince myself:
            $$
            beginvmatrix
            a_11-lambda & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=\
            =beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_22-lambda & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Now you see the problem has split into two parts. Deal with each one seperately: in the first one, move on to the next column and again split the determinant up as
            $$
            beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22 & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_11 & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Do the same for the second one. As you keep doing this, a huge number of determinants will crop up ($2^n$ in total - which is just the sum of the binomial coefficients $nchoose k$, already suggestive). You can view this as a decision tree: at each step, you say whether you want to keep the $i$-th column or not. After you pick, you won't touch the previous columns again. So doing this will give you square determinants with certain columns missing. If you pick some columns (equivalently, some diagonal elements) to remove, say "NO" to every decision where you remove one of the desired columns and "YES" otherwise. This defines a path down the binary tree which terminantes when no more $lambda$'s are left. Since every selection of $k$ diagonal elements corresponds to exactly one path down the binary tree with $n-k$ "YES"'s, you will get $(-1)^n-klambda^n-k$ in front of each such determinant and so when you sum it all up, you get precisely the sum of of all diagonal-centered $ktimes k$ determinants.



            In particular, you get the linear term by removing only $1$ column and row and summing over all possibilities. Since these are determinants of the principal minors which feature in the definition of the adjugate matrix, what you calculate is just the sum of the diagonal elements of the adjugate - that is, its trace.



            If you don't believe all this mumbo-jumbo (which is perfectly reasonable), I strongly encourage you to try a $4times 4$ case since that is only $16$ determinants in total, or, if you have a lot of paper, the $5times 5$ case. That might help you get a "feel" for the pattern.



            I also think this exact argument could be made more rigorous if you threw in some arguments like "at the $i$-th step, determinants of a certain size will appear at a certain power of $lambda$" and try doing this by induction, but I never actually went through with this so it's just a hunch at this point.




            So I tried my hand at formalizing this a bit and here's my attempt. Let $D(b_1,b_2,...,b_n)$ denote the various determinants that will appear, where $b_iin0,1,2$. $b_i=0$ means that you remove the $i$-th row and column entirely, $b_i=1$ means that you only remove the $lambda$ at $a_ii$ and $b_i=2$ means "leave the $i$-th diagonal element and its row and column as is". For example:
            $$
            D(2,2,2)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22-lambda & a_23 \
            a_31 & a_32 & a_33-lambda
            endvmatrixquad
            D(2,1,1)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22 & a_23 \
            a_31 & a_32 & a_33
            endvmatrix\
            D(2,1,0)=beginvmatrix
            a_11-lambda & a_12\
            a_21 & a_22
            endvmatrix
            $$

            We have the reduction formula: $$D(b_1,...,2,...,b_n)=D(b_1,...,1,...,b_n)-lambda D(b_1,...,0,...,b_n)$$
            We start with $D(2_1,2_2,...,2_n)$. Write $bfb$ for the vector of $b_i$ and let $Vert bVert$ be the number of $1$'s in $bfb$. If we proceed by induction, we can say that
            $$D(0,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(0,bf b)(-lambda)^n-1-Vertbf bVert$$
            Since the $0$ at the front plays no role and the reduction formula is the same regardless of what the first $b_i$ is, we must have that also:
            $$D(1,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(1,bf b)(-lambda)^n-1-Vertbf bVert$$
            This can be rigorously demonstrated by observing that if we apply the reduction formula to $D(0,bf b)$ on some index, we get the same as if we would for $D(1text or 2,bf b)$ but for the first index. By applying reduction inductively on the indices, we eliminate all $2$'s but since at each step the formula is the same except that we change the first index, after all's said and done, we're left with identical formulas but for the first index.

            Anyway, we have:
            $$
            D(2_1,2_2,...,2_n)=D(1,2,...,2)-lambda D(0,2,...,2)=\
            =sum_dimbf b=n-1\quad ,2notinbf bleft(D(1,bf b)(-lambda)^n-1-Vertbf bVert+D(0,bf b)(-lambda)^n-Vertbf bVertright)=sum_dimbf b=n\,,,,, 2notinbf bD(bf b)(-lambda)^n-Vertbf bVert
            $$

            This establishes the inductive step and completes the proof.




            Due to the helpful suggestion of @amd in the comments, here's an approach with Leibniz's formula. Here, $N=1,...,n$.
            beginalign
            &det(A-lambda I)=sum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_i=1^n(a_isigma(i)-lambdadelta_isigma(i))=\
            &=sum_sigmainmathrmSym(n)mathrmsgn(sigma)sum_Asubset Nprod_iin Aa_isigma(i)prod_jin N-A(-lambda)delta_jsigma(j)=\
            &=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_iin Aa_isigma(i)prod_jin N-Adelta_jsigma(j)
            endalign

            In the last expression we have that $prod_jin N-Adelta_jsigma(j)$ which is $0$ unless $N-A$ contains only fixed points of $sigma$ - in other words, we need only sum over $sigma$ whose fixed points are a superset of $N-A$. But these new permutations form a group which is essentially just $mathrmSym(A)$, the permutations of $A$. The sign of a permutation of length $m$ is $(-1)^k$ where $k=m-text(number of disjoint cycles that make up $sigma$)$ - when we drop the stationary points from $sigma$, $m$ decreases but the number of disjoint cycles decreases by the exact same amount since stationary points correspond to trivial cycles. Hence the sign remains the same and we have:
            $$
            det(A-lambda I)=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(A)mathrmsgn(sigma)prod_iin Aa_isigma(i)
            $$

            But the sums over $sigmainmathrmSym(A)$ are now just determinants of the principal minors and we are done.






            share|cite|improve this answer











            $endgroup$












            • $begingroup$
              I believe that the Leibniz formula for determinants can be used to prove this.
              $endgroup$
              – amd
              7 hours ago










            • $begingroup$
              Possibly, but I dread the thought of using that formula, I just find it really unwieldy.
              $endgroup$
              – J_P
              6 hours ago











            • $begingroup$
              It’s awful for computations, but great for proofs. In particular, your construction is equivalent to picking out the terms that involve exactly $n-k$ $lambda$’s.
              $endgroup$
              – amd
              6 hours ago










            • $begingroup$
              Wow, you're right! I think I managed to do it. Well, I'm not absolutely certain about my combinatorics (have to think about permutations a bit at one step) but I think it works. I'll also add this approach to my answer unless you were thinking of posting it yourself in which case tell me and I'll delete my edit.
              $endgroup$
              – J_P
              5 hours ago










            • $begingroup$
              You need to be a bit careful about contributions from terms with more $lambda$’s, but they might cancel nicely. Anyway, that formula is where I’d start looking for a proof.
              $endgroup$
              – amd
              5 hours ago
















            0












            $begingroup$

            I define $phi_A(x)=det(A-lambda I)$. You get the coefficient at $lambda^n-k$ in the following way: pick $k$ diagonal entries $a_i_1i_1,...,a_i_ki_k$. These define a $ktimes k$ minor of $A$ that you get by removing every row and column that does not contain one of the chosen diagonal elements. Calculate the determinant of this minor and sum over all $nchoose k$ selections of the $k$ diagonal entries. Up to sign (which is just alternating) that is your $lambda^n-k$ coefficient.



            I have to say that I don't know a precise formal proof (EDIT: go to the bottom of the answer, I've now added an attempt at formalizing what follows), I like to view this in a more algorithmic fashion. Here's how I convince myself:
            $$
            beginvmatrix
            a_11-lambda & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=\
            =beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_22-lambda & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Now you see the problem has split into two parts. Deal with each one seperately: in the first one, move on to the next column and again split the determinant up as
            $$
            beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22 & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_11 & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Do the same for the second one. As you keep doing this, a huge number of determinants will crop up ($2^n$ in total - which is just the sum of the binomial coefficients $nchoose k$, already suggestive). You can view this as a decision tree: at each step, you say whether you want to keep the $i$-th column or not. After you pick, you won't touch the previous columns again. So doing this will give you square determinants with certain columns missing. If you pick some columns (equivalently, some diagonal elements) to remove, say "NO" to every decision where you remove one of the desired columns and "YES" otherwise. This defines a path down the binary tree which terminantes when no more $lambda$'s are left. Since every selection of $k$ diagonal elements corresponds to exactly one path down the binary tree with $n-k$ "YES"'s, you will get $(-1)^n-klambda^n-k$ in front of each such determinant and so when you sum it all up, you get precisely the sum of of all diagonal-centered $ktimes k$ determinants.



            In particular, you get the linear term by removing only $1$ column and row and summing over all possibilities. Since these are determinants of the principal minors which feature in the definition of the adjugate matrix, what you calculate is just the sum of the diagonal elements of the adjugate - that is, its trace.



            If you don't believe all this mumbo-jumbo (which is perfectly reasonable), I strongly encourage you to try a $4times 4$ case since that is only $16$ determinants in total, or, if you have a lot of paper, the $5times 5$ case. That might help you get a "feel" for the pattern.



            I also think this exact argument could be made more rigorous if you threw in some arguments like "at the $i$-th step, determinants of a certain size will appear at a certain power of $lambda$" and try doing this by induction, but I never actually went through with this so it's just a hunch at this point.




            So I tried my hand at formalizing this a bit and here's my attempt. Let $D(b_1,b_2,...,b_n)$ denote the various determinants that will appear, where $b_iin0,1,2$. $b_i=0$ means that you remove the $i$-th row and column entirely, $b_i=1$ means that you only remove the $lambda$ at $a_ii$ and $b_i=2$ means "leave the $i$-th diagonal element and its row and column as is". For example:
            $$
            D(2,2,2)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22-lambda & a_23 \
            a_31 & a_32 & a_33-lambda
            endvmatrixquad
            D(2,1,1)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22 & a_23 \
            a_31 & a_32 & a_33
            endvmatrix\
            D(2,1,0)=beginvmatrix
            a_11-lambda & a_12\
            a_21 & a_22
            endvmatrix
            $$

            We have the reduction formula: $$D(b_1,...,2,...,b_n)=D(b_1,...,1,...,b_n)-lambda D(b_1,...,0,...,b_n)$$
            We start with $D(2_1,2_2,...,2_n)$. Write $bfb$ for the vector of $b_i$ and let $Vert bVert$ be the number of $1$'s in $bfb$. If we proceed by induction, we can say that
            $$D(0,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(0,bf b)(-lambda)^n-1-Vertbf bVert$$
            Since the $0$ at the front plays no role and the reduction formula is the same regardless of what the first $b_i$ is, we must have that also:
            $$D(1,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(1,bf b)(-lambda)^n-1-Vertbf bVert$$
            This can be rigorously demonstrated by observing that if we apply the reduction formula to $D(0,bf b)$ on some index, we get the same as if we would for $D(1text or 2,bf b)$ but for the first index. By applying reduction inductively on the indices, we eliminate all $2$'s but since at each step the formula is the same except that we change the first index, after all's said and done, we're left with identical formulas but for the first index.

            Anyway, we have:
            $$
            D(2_1,2_2,...,2_n)=D(1,2,...,2)-lambda D(0,2,...,2)=\
            =sum_dimbf b=n-1\quad ,2notinbf bleft(D(1,bf b)(-lambda)^n-1-Vertbf bVert+D(0,bf b)(-lambda)^n-Vertbf bVertright)=sum_dimbf b=n\,,,,, 2notinbf bD(bf b)(-lambda)^n-Vertbf bVert
            $$

            This establishes the inductive step and completes the proof.




            Due to the helpful suggestion of @amd in the comments, here's an approach with Leibniz's formula. Here, $N=1,...,n$.
            beginalign
            &det(A-lambda I)=sum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_i=1^n(a_isigma(i)-lambdadelta_isigma(i))=\
            &=sum_sigmainmathrmSym(n)mathrmsgn(sigma)sum_Asubset Nprod_iin Aa_isigma(i)prod_jin N-A(-lambda)delta_jsigma(j)=\
            &=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_iin Aa_isigma(i)prod_jin N-Adelta_jsigma(j)
            endalign

            In the last expression we have that $prod_jin N-Adelta_jsigma(j)$ which is $0$ unless $N-A$ contains only fixed points of $sigma$ - in other words, we need only sum over $sigma$ whose fixed points are a superset of $N-A$. But these new permutations form a group which is essentially just $mathrmSym(A)$, the permutations of $A$. The sign of a permutation of length $m$ is $(-1)^k$ where $k=m-text(number of disjoint cycles that make up $sigma$)$ - when we drop the stationary points from $sigma$, $m$ decreases but the number of disjoint cycles decreases by the exact same amount since stationary points correspond to trivial cycles. Hence the sign remains the same and we have:
            $$
            det(A-lambda I)=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(A)mathrmsgn(sigma)prod_iin Aa_isigma(i)
            $$

            But the sums over $sigmainmathrmSym(A)$ are now just determinants of the principal minors and we are done.






            share|cite|improve this answer











            $endgroup$












            • $begingroup$
              I believe that the Leibniz formula for determinants can be used to prove this.
              $endgroup$
              – amd
              7 hours ago










            • $begingroup$
              Possibly, but I dread the thought of using that formula, I just find it really unwieldy.
              $endgroup$
              – J_P
              6 hours ago











            • $begingroup$
              It’s awful for computations, but great for proofs. In particular, your construction is equivalent to picking out the terms that involve exactly $n-k$ $lambda$’s.
              $endgroup$
              – amd
              6 hours ago










            • $begingroup$
              Wow, you're right! I think I managed to do it. Well, I'm not absolutely certain about my combinatorics (have to think about permutations a bit at one step) but I think it works. I'll also add this approach to my answer unless you were thinking of posting it yourself in which case tell me and I'll delete my edit.
              $endgroup$
              – J_P
              5 hours ago










            • $begingroup$
              You need to be a bit careful about contributions from terms with more $lambda$’s, but they might cancel nicely. Anyway, that formula is where I’d start looking for a proof.
              $endgroup$
              – amd
              5 hours ago














            0












            0








            0





            $begingroup$

            I define $phi_A(x)=det(A-lambda I)$. You get the coefficient at $lambda^n-k$ in the following way: pick $k$ diagonal entries $a_i_1i_1,...,a_i_ki_k$. These define a $ktimes k$ minor of $A$ that you get by removing every row and column that does not contain one of the chosen diagonal elements. Calculate the determinant of this minor and sum over all $nchoose k$ selections of the $k$ diagonal entries. Up to sign (which is just alternating) that is your $lambda^n-k$ coefficient.



            I have to say that I don't know a precise formal proof (EDIT: go to the bottom of the answer, I've now added an attempt at formalizing what follows), I like to view this in a more algorithmic fashion. Here's how I convince myself:
            $$
            beginvmatrix
            a_11-lambda & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=\
            =beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_22-lambda & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Now you see the problem has split into two parts. Deal with each one seperately: in the first one, move on to the next column and again split the determinant up as
            $$
            beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22 & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_11 & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Do the same for the second one. As you keep doing this, a huge number of determinants will crop up ($2^n$ in total - which is just the sum of the binomial coefficients $nchoose k$, already suggestive). You can view this as a decision tree: at each step, you say whether you want to keep the $i$-th column or not. After you pick, you won't touch the previous columns again. So doing this will give you square determinants with certain columns missing. If you pick some columns (equivalently, some diagonal elements) to remove, say "NO" to every decision where you remove one of the desired columns and "YES" otherwise. This defines a path down the binary tree which terminantes when no more $lambda$'s are left. Since every selection of $k$ diagonal elements corresponds to exactly one path down the binary tree with $n-k$ "YES"'s, you will get $(-1)^n-klambda^n-k$ in front of each such determinant and so when you sum it all up, you get precisely the sum of of all diagonal-centered $ktimes k$ determinants.



            In particular, you get the linear term by removing only $1$ column and row and summing over all possibilities. Since these are determinants of the principal minors which feature in the definition of the adjugate matrix, what you calculate is just the sum of the diagonal elements of the adjugate - that is, its trace.



            If you don't believe all this mumbo-jumbo (which is perfectly reasonable), I strongly encourage you to try a $4times 4$ case since that is only $16$ determinants in total, or, if you have a lot of paper, the $5times 5$ case. That might help you get a "feel" for the pattern.



            I also think this exact argument could be made more rigorous if you threw in some arguments like "at the $i$-th step, determinants of a certain size will appear at a certain power of $lambda$" and try doing this by induction, but I never actually went through with this so it's just a hunch at this point.




            So I tried my hand at formalizing this a bit and here's my attempt. Let $D(b_1,b_2,...,b_n)$ denote the various determinants that will appear, where $b_iin0,1,2$. $b_i=0$ means that you remove the $i$-th row and column entirely, $b_i=1$ means that you only remove the $lambda$ at $a_ii$ and $b_i=2$ means "leave the $i$-th diagonal element and its row and column as is". For example:
            $$
            D(2,2,2)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22-lambda & a_23 \
            a_31 & a_32 & a_33-lambda
            endvmatrixquad
            D(2,1,1)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22 & a_23 \
            a_31 & a_32 & a_33
            endvmatrix\
            D(2,1,0)=beginvmatrix
            a_11-lambda & a_12\
            a_21 & a_22
            endvmatrix
            $$

            We have the reduction formula: $$D(b_1,...,2,...,b_n)=D(b_1,...,1,...,b_n)-lambda D(b_1,...,0,...,b_n)$$
            We start with $D(2_1,2_2,...,2_n)$. Write $bfb$ for the vector of $b_i$ and let $Vert bVert$ be the number of $1$'s in $bfb$. If we proceed by induction, we can say that
            $$D(0,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(0,bf b)(-lambda)^n-1-Vertbf bVert$$
            Since the $0$ at the front plays no role and the reduction formula is the same regardless of what the first $b_i$ is, we must have that also:
            $$D(1,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(1,bf b)(-lambda)^n-1-Vertbf bVert$$
            This can be rigorously demonstrated by observing that if we apply the reduction formula to $D(0,bf b)$ on some index, we get the same as if we would for $D(1text or 2,bf b)$ but for the first index. By applying reduction inductively on the indices, we eliminate all $2$'s but since at each step the formula is the same except that we change the first index, after all's said and done, we're left with identical formulas but for the first index.

            Anyway, we have:
            $$
            D(2_1,2_2,...,2_n)=D(1,2,...,2)-lambda D(0,2,...,2)=\
            =sum_dimbf b=n-1\quad ,2notinbf bleft(D(1,bf b)(-lambda)^n-1-Vertbf bVert+D(0,bf b)(-lambda)^n-Vertbf bVertright)=sum_dimbf b=n\,,,,, 2notinbf bD(bf b)(-lambda)^n-Vertbf bVert
            $$

            This establishes the inductive step and completes the proof.




            Due to the helpful suggestion of @amd in the comments, here's an approach with Leibniz's formula. Here, $N=1,...,n$.
            beginalign
            &det(A-lambda I)=sum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_i=1^n(a_isigma(i)-lambdadelta_isigma(i))=\
            &=sum_sigmainmathrmSym(n)mathrmsgn(sigma)sum_Asubset Nprod_iin Aa_isigma(i)prod_jin N-A(-lambda)delta_jsigma(j)=\
            &=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_iin Aa_isigma(i)prod_jin N-Adelta_jsigma(j)
            endalign

            In the last expression we have that $prod_jin N-Adelta_jsigma(j)$ which is $0$ unless $N-A$ contains only fixed points of $sigma$ - in other words, we need only sum over $sigma$ whose fixed points are a superset of $N-A$. But these new permutations form a group which is essentially just $mathrmSym(A)$, the permutations of $A$. The sign of a permutation of length $m$ is $(-1)^k$ where $k=m-text(number of disjoint cycles that make up $sigma$)$ - when we drop the stationary points from $sigma$, $m$ decreases but the number of disjoint cycles decreases by the exact same amount since stationary points correspond to trivial cycles. Hence the sign remains the same and we have:
            $$
            det(A-lambda I)=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(A)mathrmsgn(sigma)prod_iin Aa_isigma(i)
            $$

            But the sums over $sigmainmathrmSym(A)$ are now just determinants of the principal minors and we are done.






            share|cite|improve this answer











            $endgroup$



            I define $phi_A(x)=det(A-lambda I)$. You get the coefficient at $lambda^n-k$ in the following way: pick $k$ diagonal entries $a_i_1i_1,...,a_i_ki_k$. These define a $ktimes k$ minor of $A$ that you get by removing every row and column that does not contain one of the chosen diagonal elements. Calculate the determinant of this minor and sum over all $nchoose k$ selections of the $k$ diagonal entries. Up to sign (which is just alternating) that is your $lambda^n-k$ coefficient.



            I have to say that I don't know a precise formal proof (EDIT: go to the bottom of the answer, I've now added an attempt at formalizing what follows), I like to view this in a more algorithmic fashion. Here's how I convince myself:
            $$
            beginvmatrix
            a_11-lambda & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=\
            =beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_22-lambda & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Now you see the problem has split into two parts. Deal with each one seperately: in the first one, move on to the next column and again split the determinant up as
            $$
            beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22-lambda & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix=beginvmatrix
            a_11 & a_12 & ... & a_1n \
            a_21 & a_22 & ... & a_2n \
            ... & ... & ... & ... \
            a_n1 & a_n2 & ...& a_nn-lambda
            endvmatrix-lambdabeginvmatrix
            a_11 & ... & a_2n \
            ... & ... & ... \
            a_n2 & ...& a_nn-lambda
            endvmatrix
            $$

            Do the same for the second one. As you keep doing this, a huge number of determinants will crop up ($2^n$ in total - which is just the sum of the binomial coefficients $nchoose k$, already suggestive). You can view this as a decision tree: at each step, you say whether you want to keep the $i$-th column or not. After you pick, you won't touch the previous columns again. So doing this will give you square determinants with certain columns missing. If you pick some columns (equivalently, some diagonal elements) to remove, say "NO" to every decision where you remove one of the desired columns and "YES" otherwise. This defines a path down the binary tree which terminantes when no more $lambda$'s are left. Since every selection of $k$ diagonal elements corresponds to exactly one path down the binary tree with $n-k$ "YES"'s, you will get $(-1)^n-klambda^n-k$ in front of each such determinant and so when you sum it all up, you get precisely the sum of of all diagonal-centered $ktimes k$ determinants.



            In particular, you get the linear term by removing only $1$ column and row and summing over all possibilities. Since these are determinants of the principal minors which feature in the definition of the adjugate matrix, what you calculate is just the sum of the diagonal elements of the adjugate - that is, its trace.



            If you don't believe all this mumbo-jumbo (which is perfectly reasonable), I strongly encourage you to try a $4times 4$ case since that is only $16$ determinants in total, or, if you have a lot of paper, the $5times 5$ case. That might help you get a "feel" for the pattern.



            I also think this exact argument could be made more rigorous if you threw in some arguments like "at the $i$-th step, determinants of a certain size will appear at a certain power of $lambda$" and try doing this by induction, but I never actually went through with this so it's just a hunch at this point.




            So I tried my hand at formalizing this a bit and here's my attempt. Let $D(b_1,b_2,...,b_n)$ denote the various determinants that will appear, where $b_iin0,1,2$. $b_i=0$ means that you remove the $i$-th row and column entirely, $b_i=1$ means that you only remove the $lambda$ at $a_ii$ and $b_i=2$ means "leave the $i$-th diagonal element and its row and column as is". For example:
            $$
            D(2,2,2)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22-lambda & a_23 \
            a_31 & a_32 & a_33-lambda
            endvmatrixquad
            D(2,1,1)=beginvmatrix
            a_11-lambda & a_12 & a_13 \
            a_21 & a_22 & a_23 \
            a_31 & a_32 & a_33
            endvmatrix\
            D(2,1,0)=beginvmatrix
            a_11-lambda & a_12\
            a_21 & a_22
            endvmatrix
            $$

            We have the reduction formula: $$D(b_1,...,2,...,b_n)=D(b_1,...,1,...,b_n)-lambda D(b_1,...,0,...,b_n)$$
            We start with $D(2_1,2_2,...,2_n)$. Write $bfb$ for the vector of $b_i$ and let $Vert bVert$ be the number of $1$'s in $bfb$. If we proceed by induction, we can say that
            $$D(0,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(0,bf b)(-lambda)^n-1-Vertbf bVert$$
            Since the $0$ at the front plays no role and the reduction formula is the same regardless of what the first $b_i$ is, we must have that also:
            $$D(1,2_1,...,2_n-1)=sum_dimbf b=n-1\quad ,2notinbf bD(1,bf b)(-lambda)^n-1-Vertbf bVert$$
            This can be rigorously demonstrated by observing that if we apply the reduction formula to $D(0,bf b)$ on some index, we get the same as if we would for $D(1text or 2,bf b)$ but for the first index. By applying reduction inductively on the indices, we eliminate all $2$'s but since at each step the formula is the same except that we change the first index, after all's said and done, we're left with identical formulas but for the first index.

            Anyway, we have:
            $$
            D(2_1,2_2,...,2_n)=D(1,2,...,2)-lambda D(0,2,...,2)=\
            =sum_dimbf b=n-1\quad ,2notinbf bleft(D(1,bf b)(-lambda)^n-1-Vertbf bVert+D(0,bf b)(-lambda)^n-Vertbf bVertright)=sum_dimbf b=n\,,,,, 2notinbf bD(bf b)(-lambda)^n-Vertbf bVert
            $$

            This establishes the inductive step and completes the proof.




            Due to the helpful suggestion of @amd in the comments, here's an approach with Leibniz's formula. Here, $N=1,...,n$.
            beginalign
            &det(A-lambda I)=sum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_i=1^n(a_isigma(i)-lambdadelta_isigma(i))=\
            &=sum_sigmainmathrmSym(n)mathrmsgn(sigma)sum_Asubset Nprod_iin Aa_isigma(i)prod_jin N-A(-lambda)delta_jsigma(j)=\
            &=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(n)mathrmsgn(sigma)prod_iin Aa_isigma(i)prod_jin N-Adelta_jsigma(j)
            endalign

            In the last expression we have that $prod_jin N-Adelta_jsigma(j)$ which is $0$ unless $N-A$ contains only fixed points of $sigma$ - in other words, we need only sum over $sigma$ whose fixed points are a superset of $N-A$. But these new permutations form a group which is essentially just $mathrmSym(A)$, the permutations of $A$. The sign of a permutation of length $m$ is $(-1)^k$ where $k=m-text(number of disjoint cycles that make up $sigma$)$ - when we drop the stationary points from $sigma$, $m$ decreases but the number of disjoint cycles decreases by the exact same amount since stationary points correspond to trivial cycles. Hence the sign remains the same and we have:
            $$
            det(A-lambda I)=sum_Asubset N(-lambda)^n-Vert AVertsum_sigmainmathrmSym(A)mathrmsgn(sigma)prod_iin Aa_isigma(i)
            $$

            But the sums over $sigmainmathrmSym(A)$ are now just determinants of the principal minors and we are done.







            share|cite|improve this answer














            share|cite|improve this answer



            share|cite|improve this answer








            edited 5 hours ago

























            answered 8 hours ago









            J_PJ_P

            1,0941 silver badge12 bronze badges




            1,0941 silver badge12 bronze badges











            • $begingroup$
              I believe that the Leibniz formula for determinants can be used to prove this.
              $endgroup$
              – amd
              7 hours ago










            • $begingroup$
              Possibly, but I dread the thought of using that formula, I just find it really unwieldy.
              $endgroup$
              – J_P
              6 hours ago











            • $begingroup$
              It’s awful for computations, but great for proofs. In particular, your construction is equivalent to picking out the terms that involve exactly $n-k$ $lambda$’s.
              $endgroup$
              – amd
              6 hours ago










            • $begingroup$
              Wow, you're right! I think I managed to do it. Well, I'm not absolutely certain about my combinatorics (have to think about permutations a bit at one step) but I think it works. I'll also add this approach to my answer unless you were thinking of posting it yourself in which case tell me and I'll delete my edit.
              $endgroup$
              – J_P
              5 hours ago










            • $begingroup$
              You need to be a bit careful about contributions from terms with more $lambda$’s, but they might cancel nicely. Anyway, that formula is where I’d start looking for a proof.
              $endgroup$
              – amd
              5 hours ago

















            • $begingroup$
              I believe that the Leibniz formula for determinants can be used to prove this.
              $endgroup$
              – amd
              7 hours ago










            • $begingroup$
              Possibly, but I dread the thought of using that formula, I just find it really unwieldy.
              $endgroup$
              – J_P
              6 hours ago











            • $begingroup$
              It’s awful for computations, but great for proofs. In particular, your construction is equivalent to picking out the terms that involve exactly $n-k$ $lambda$’s.
              $endgroup$
              – amd
              6 hours ago










            • $begingroup$
              Wow, you're right! I think I managed to do it. Well, I'm not absolutely certain about my combinatorics (have to think about permutations a bit at one step) but I think it works. I'll also add this approach to my answer unless you were thinking of posting it yourself in which case tell me and I'll delete my edit.
              $endgroup$
              – J_P
              5 hours ago










            • $begingroup$
              You need to be a bit careful about contributions from terms with more $lambda$’s, but they might cancel nicely. Anyway, that formula is where I’d start looking for a proof.
              $endgroup$
              – amd
              5 hours ago
















            $begingroup$
            I believe that the Leibniz formula for determinants can be used to prove this.
            $endgroup$
            – amd
            7 hours ago




            $begingroup$
            I believe that the Leibniz formula for determinants can be used to prove this.
            $endgroup$
            – amd
            7 hours ago












            $begingroup$
            Possibly, but I dread the thought of using that formula, I just find it really unwieldy.
            $endgroup$
            – J_P
            6 hours ago





            $begingroup$
            Possibly, but I dread the thought of using that formula, I just find it really unwieldy.
            $endgroup$
            – J_P
            6 hours ago













            $begingroup$
            It’s awful for computations, but great for proofs. In particular, your construction is equivalent to picking out the terms that involve exactly $n-k$ $lambda$’s.
            $endgroup$
            – amd
            6 hours ago




            $begingroup$
            It’s awful for computations, but great for proofs. In particular, your construction is equivalent to picking out the terms that involve exactly $n-k$ $lambda$’s.
            $endgroup$
            – amd
            6 hours ago












            $begingroup$
            Wow, you're right! I think I managed to do it. Well, I'm not absolutely certain about my combinatorics (have to think about permutations a bit at one step) but I think it works. I'll also add this approach to my answer unless you were thinking of posting it yourself in which case tell me and I'll delete my edit.
            $endgroup$
            – J_P
            5 hours ago




            $begingroup$
            Wow, you're right! I think I managed to do it. Well, I'm not absolutely certain about my combinatorics (have to think about permutations a bit at one step) but I think it works. I'll also add this approach to my answer unless you were thinking of posting it yourself in which case tell me and I'll delete my edit.
            $endgroup$
            – J_P
            5 hours ago












            $begingroup$
            You need to be a bit careful about contributions from terms with more $lambda$’s, but they might cancel nicely. Anyway, that formula is where I’d start looking for a proof.
            $endgroup$
            – amd
            5 hours ago





            $begingroup$
            You need to be a bit careful about contributions from terms with more $lambda$’s, but they might cancel nicely. Anyway, that formula is where I’d start looking for a proof.
            $endgroup$
            – amd
            5 hours ago


















            draft saved

            draft discarded
















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid


            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.

            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3272063%2fcoefficients-of-the-characteristic-polynomial%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            CRJq esWuELKzyY24uEO2 hIBSrh9,XsJ4 skNfu Py4n6Igf hQXs6kN7fGM1tOgrCXi,DI9JFOTrSx
            UuyUSX6dUxkls,sazM28

            Popular posts from this blog

            19. јануар Садржај Догађаји Рођења Смрти Празници и дани сећања Види још Референце Мени за навигацијуу

            Israel Cuprins Etimologie | Istorie | Geografie | Politică | Demografie | Educație | Economie | Cultură | Note explicative | Note bibliografice | Bibliografie | Legături externe | Meniu de navigaresite web oficialfacebooktweeterGoogle+Instagramcanal YouTubeInstagramtextmodificaremodificarewww.technion.ac.ilnew.huji.ac.ilwww.weizmann.ac.ilwww1.biu.ac.ilenglish.tau.ac.ilwww.haifa.ac.ilin.bgu.ac.ilwww.openu.ac.ilwww.ariel.ac.ilCIA FactbookHarta Israelului"Negotiating Jerusalem," Palestine–Israel JournalThe Schizoid Nature of Modern Hebrew: A Slavic Language in Search of a Semitic Past„Arabic in Israel: an official language and a cultural bridge”„Latest Population Statistics for Israel”„Israel Population”„Tables”„Report for Selected Countries and Subjects”Human Development Report 2016: Human Development for Everyone„Distribution of family income - Gini index”The World FactbookJerusalem Law„Israel”„Israel”„Zionist Leaders: David Ben-Gurion 1886–1973”„The status of Jerusalem”„Analysis: Kadima's big plans”„Israel's Hard-Learned Lessons”„The Legacy of Undefined Borders, Tel Aviv Notes No. 40, 5 iunie 2002”„Israel Journal: A Land Without Borders”„Population”„Israel closes decade with population of 7.5 million”Time Series-DataBank„Selected Statistics on Jerusalem Day 2007 (Hebrew)”Golan belongs to Syria, Druze protestGlobal Survey 2006: Middle East Progress Amid Global Gains in FreedomWHO: Life expectancy in Israel among highest in the worldInternational Monetary Fund, World Economic Outlook Database, April 2011: Nominal GDP list of countries. Data for the year 2010.„Israel's accession to the OECD”Popular Opinion„On the Move”Hosea 12:5„Walking the Bible Timeline”„Palestine: History”„Return to Zion”An invention called 'the Jewish people' – Haaretz – Israel NewsoriginalJewish and Non-Jewish Population of Palestine-Israel (1517–2004)ImmigrationJewishvirtuallibrary.orgChapter One: The Heralders of Zionism„The birth of modern Israel: A scrap of paper that changed history”„League of Nations: The Mandate for Palestine, 24 iulie 1922”The Population of Palestine Prior to 1948originalBackground Paper No. 47 (ST/DPI/SER.A/47)History: Foreign DominationTwo Hundred and Seventh Plenary Meeting„Israel (Labor Zionism)”Population, by Religion and Population GroupThe Suez CrisisAdolf EichmannJustice Ministry Reply to Amnesty International Report„The Interregnum”Israel Ministry of Foreign Affairs – The Palestinian National Covenant- July 1968Research on terrorism: trends, achievements & failuresThe Routledge Atlas of the Arab–Israeli conflict: The Complete History of the Struggle and the Efforts to Resolve It"George Habash, Palestinian Terrorism Tactician, Dies at 82."„1973: Arab states attack Israeli forces”Agranat Commission„Has Israel Annexed East Jerusalem?”original„After 4 Years, Intifada Still Smolders”From the End of the Cold War to 2001originalThe Oslo Accords, 1993Israel-PLO Recognition – Exchange of Letters between PM Rabin and Chairman Arafat – Sept 9- 1993Foundation for Middle East PeaceSources of Population Growth: Total Israeli Population and Settler Population, 1991–2003original„Israel marks Rabin assassination”The Wye River Memorandumoriginal„West Bank barrier route disputed, Israeli missile kills 2”"Permanent Ceasefire to Be Based on Creation Of Buffer Zone Free of Armed Personnel Other than UN, Lebanese Forces"„Hezbollah kills 8 soldiers, kidnaps two in offensive on northern border”„Olmert confirms peace talks with Syria”„Battleground Gaza: Israeli ground forces invade the strip”„IDF begins Gaza troop withdrawal, hours after ending 3-week offensive”„THE LAND: Geography and Climate”„Area of districts, sub-districts, natural regions and lakes”„Israel - Geography”„Makhteshim Country”Israel and the Palestinian Territories„Makhtesh Ramon”„The Living Dead Sea”„Temperatures reach record high in Pakistan”„Climate Extremes In Israel”Israel in figures„Deuteronom”„JNF: 240 million trees planted since 1901”„Vegetation of Israel and Neighboring Countries”Environmental Law in Israel„Executive branch”„Israel's election process explained”„The Electoral System in Israel”„Constitution for Israel”„All 120 incoming Knesset members”„Statul ISRAEL”„The Judiciary: The Court System”„Israel's high court unique in region”„Israel and the International Criminal Court: A Legal Battlefield”„Localities and population, by population group, district, sub-district and natural region”„Israel: Districts, Major Cities, Urban Localities & Metropolitan Areas”„Israel-Egypt Relations: Background & Overview of Peace Treaty”„Solana to Haaretz: New Rules of War Needed for Age of Terror”„Israel's Announcement Regarding Settlements”„United Nations Security Council Resolution 497”„Security Council resolution 478 (1980) on the status of Jerusalem”„Arabs will ask U.N. to seek razing of Israeli wall”„Olmert: Willing to trade land for peace”„Mapping Peace between Syria and Israel”„Egypt: Israel must accept the land-for-peace formula”„Israel: Age structure from 2005 to 2015”„Global, regional, and national disability-adjusted life years (DALYs) for 306 diseases and injuries and healthy life expectancy (HALE) for 188 countries, 1990–2013: quantifying the epidemiological transition”10.1016/S0140-6736(15)61340-X„World Health Statistics 2014”„Life expectancy for Israeli men world's 4th highest”„Family Structure and Well-Being Across Israel's Diverse Population”„Fertility among Jewish and Muslim Women in Israel, by Level of Religiosity, 1979-2009”„Israel leaders in birth rate, but poverty major challenge”„Ethnic Groups”„Israel's population: Over 8.5 million”„Israel - Ethnic groups”„Jews, by country of origin and age”„Minority Communities in Israel: Background & Overview”„Israel”„Language in Israel”„Selected Data from the 2011 Social Survey on Mastery of the Hebrew Language and Usage of Languages”„Religions”„5 facts about Israeli Druze, a unique religious and ethnic group”„Israël”Israel Country Study Guide„Haredi city in Negev – blessing or curse?”„New town Harish harbors hopes of being more than another Pleasantville”„List of localities, in alphabetical order”„Muncitorii români, doriți în Israel”„Prietenia româno-israeliană la nevoie se cunoaște”„The Higher Education System in Israel”„Middle East”„Academic Ranking of World Universities 2016”„Israel”„Israel”„Jewish Nobel Prize Winners”„All Nobel Prizes in Literature”„All Nobel Peace Prizes”„All Prizes in Economic Sciences”„All Nobel Prizes in Chemistry”„List of Fields Medallists”„Sakharov Prize”„Țara care și-a sfidat "destinul" și se bate umăr la umăr cu Silicon Valley”„Apple's R&D center in Israel grew to about 800 employees”„Tim Cook: Apple's Herzliya R&D center second-largest in world”„Lecții de economie de la Israel”„Land use”Israel Investment and Business GuideA Country Study: IsraelCentral Bureau of StatisticsFlorin Diaconu, „Kadima: Flexibilitate și pragmatism, dar nici un compromis în chestiuni vitale", în Revista Institutului Diplomatic Român, anul I, numărul I, semestrul I, 2006, pp. 71-72Florin Diaconu, „Likud: Dreapta israeliană constant opusă retrocedării teritoriilor cureite prin luptă în 1967", în Revista Institutului Diplomatic Român, anul I, numărul I, semestrul I, 2006, pp. 73-74MassadaIsraelul a crescut in 50 de ani cât alte state intr-un mileniuIsrael Government PortalIsraelIsraelIsraelmmmmmXX451232cb118646298(data)4027808-634110000 0004 0372 0767n7900328503691455-bb46-37e3-91d2-cb064a35ffcc1003570400564274ge1294033523775214929302638955X146498911146498911

            Disable console in Battlefield 1Is it possible to re-map or disable the key that brings up the console?Can't complete Battlefield 1 instalationLocational & headshot damage in Battlefield 1How do medals work in Battlefield 1?How to equip skins to your weapon in Battlefield 1Why don't my settings and single player progress get saved?How to maximize damage to a tank in Battlefield 1?Battlefield 1 vehicle position iconsHow do you un-track a medal in Battlefield 1Fort Vaux “zombie” screams and sounds - Battlefield 1How to differentiate enemies from allies in Battlefield 1 for a color-blind?