Are these square matrices always diagonalisable?Conditions for diagonalizability of $ntimes n$ anti-diagonal matricesFinding eigenvalues/vectors of a matrix and proving it is not diagonalisable.Finding if a linear transformation is diagonalisableThe diagonalisation of the two matricesIf $A in K^n times n$ is diagonalisable, the dimension of the subspace of its commuting matrices is $geq n$ nCalculating the eigenvalues of a diagonalisable linear operator $L$.How to find eigenvalues for mod 2 field?Find for which real parameter a matrix is diagonalisableSquare Roots of a Matrix: Diagonalisable Solutions.eigenvalues and eigenvectors of Diagonalisable matrices
Sci-fi novel series with instant travel between planets through gates. A river runs through the gates
Unexpected email from Yorkshire Bank
Meaning of Bloch representation
Will a top journal at least read my introduction?
Was there a shared-world project before "Thieves World"?
Packing rectangles: Does rotation ever help?
Controversial area of mathematics
How can I practically buy stocks?
Is the 5 MB static resource size limit 5,242,880 bytes or 5,000,000 bytes?
Examples of non trivial equivalence relations , I mean equivalence relations without the expression " same ... as" in their definition?
Why was the Spitfire's elliptical wing almost uncopied by other aircraft of World War 2?
Reducing vertical space in stackrel
Will tsunami waves travel forever if there was no land?
Stop and Take a Breath!
What are the potential pitfalls when using metals as a currency?
How can I place the product on a social media post better?
Can someone publish a story that happened to you?
Using a Lyapunov function to classify stability and sketching a phase portrait
Is there any limitation with Arduino Nano serial communication distance?
Was there a Viking Exchange as well as a Columbian one?
How do I use proper grammar in the negation of "have not" for the following sentence translation?
Is there a way to get a compiler for the original B programming language?
How to make a pipeline wait for end-of-file or stop after an error?
What is the relationship between spectral sequences and obstruction theory?
Are these square matrices always diagonalisable?
Conditions for diagonalizability of $ntimes n$ anti-diagonal matricesFinding eigenvalues/vectors of a matrix and proving it is not diagonalisable.Finding if a linear transformation is diagonalisableThe diagonalisation of the two matricesIf $A in K^n times n$ is diagonalisable, the dimension of the subspace of its commuting matrices is $geq n$ nCalculating the eigenvalues of a diagonalisable linear operator $L$.How to find eigenvalues for mod 2 field?Find for which real parameter a matrix is diagonalisableSquare Roots of a Matrix: Diagonalisable Solutions.eigenvalues and eigenvectors of Diagonalisable matrices
$begingroup$
When trying to solve a physics problem on decoupling a system of ODEs, I found myself needing to address the following problem:
Let $A_nin M_n(mathbb R)$ be the matrix with all $1$s above its main diagonal, all $-1$s below its diagonal, and $0$s everywhere else. Is $A_n$ always diagonalisable? If so, what is its diagonalisation (equivalently: what are its eigenvalues and corresponding eigenvectors)?
For example,
$$A_3=beginbmatrix0&1&0\-1&0&1\0&-1&0endbmatrix,quad A_5beginbmatrix0&1&0&0&0\-1&0&1&0&0\0&-1&0&1&0\0&0&-1&0&1\0&0&0&-1&0endbmatrix.$$
Assuming my code is correct, Mathematica has been able to verify that $A_n$ is always diagonalisable up to $n=1000$. If we use $chi_n(t)inmathbb Z[t]$ to denote the characteristic polynomial of $A_n$, a straightforward evaluation also shows that
$$chi_n(t)=-tchi_n-1(t)+chi_n-2(t)tag1$$
for all $ngeq4$. Furthermore, note that $A_n=-A_n^t$ so that, in the case where the dimension is even,
$$det(A_2n-lambda I)=det(A_2n^t-lambda I)=det(-A_2n-lambda I)=det(A_2n+lambda I).$$
This implies that whenever $lambda$ is an eigenvalue of $A_2n$, so is $-lambda$. In other words, $chi_2n(t)$ is always of the form $(t^2-lambda _1^2)(t^2-lambda_2^2)dotsm(t^2-lambda_n^2)$ for some $lambda_i$.
And this is where I am stuck. In order for $A_n$ to be diagonalisable, we must have that all the eigenvalues are distinct, but trying to use the recurrence $(1)$ and strong induction, or trying to use the formula for the even case have not helped at all. It seems like the most probable line of attack would be to somehow show that
$$chi_2n'(t)=2tsum_k=1^nfracchi_2n(t)t^2-lambda_k^2$$
never shares a common zero with $chi_2n$ (which would resolve the even case), though I don't see how to make this work.
Note: I do not have any clue how to actually find the eigenvalues/eigenvectors even in the case where the $A_n$ are diagonalisable. As such even if someone cannot answer the second part of the question, but can prove that the $A_n$ are diagonalisable, I would appreciate that as an answer as well. Above I tried to look at the special case where the dimension is even, though of course the proof for all odd and even $n$ is more valuable. Even if this is not possible, for my purposes I just need an unbounded subset $Ssubseteqmathbb Z$ for which the conclusion is proven for $nin S$, so any such approach is welcome too.
Thank you in advance!
linear-algebra eigenvalues-eigenvectors determinant diagonalization
$endgroup$
add a comment |
$begingroup$
When trying to solve a physics problem on decoupling a system of ODEs, I found myself needing to address the following problem:
Let $A_nin M_n(mathbb R)$ be the matrix with all $1$s above its main diagonal, all $-1$s below its diagonal, and $0$s everywhere else. Is $A_n$ always diagonalisable? If so, what is its diagonalisation (equivalently: what are its eigenvalues and corresponding eigenvectors)?
For example,
$$A_3=beginbmatrix0&1&0\-1&0&1\0&-1&0endbmatrix,quad A_5beginbmatrix0&1&0&0&0\-1&0&1&0&0\0&-1&0&1&0\0&0&-1&0&1\0&0&0&-1&0endbmatrix.$$
Assuming my code is correct, Mathematica has been able to verify that $A_n$ is always diagonalisable up to $n=1000$. If we use $chi_n(t)inmathbb Z[t]$ to denote the characteristic polynomial of $A_n$, a straightforward evaluation also shows that
$$chi_n(t)=-tchi_n-1(t)+chi_n-2(t)tag1$$
for all $ngeq4$. Furthermore, note that $A_n=-A_n^t$ so that, in the case where the dimension is even,
$$det(A_2n-lambda I)=det(A_2n^t-lambda I)=det(-A_2n-lambda I)=det(A_2n+lambda I).$$
This implies that whenever $lambda$ is an eigenvalue of $A_2n$, so is $-lambda$. In other words, $chi_2n(t)$ is always of the form $(t^2-lambda _1^2)(t^2-lambda_2^2)dotsm(t^2-lambda_n^2)$ for some $lambda_i$.
And this is where I am stuck. In order for $A_n$ to be diagonalisable, we must have that all the eigenvalues are distinct, but trying to use the recurrence $(1)$ and strong induction, or trying to use the formula for the even case have not helped at all. It seems like the most probable line of attack would be to somehow show that
$$chi_2n'(t)=2tsum_k=1^nfracchi_2n(t)t^2-lambda_k^2$$
never shares a common zero with $chi_2n$ (which would resolve the even case), though I don't see how to make this work.
Note: I do not have any clue how to actually find the eigenvalues/eigenvectors even in the case where the $A_n$ are diagonalisable. As such even if someone cannot answer the second part of the question, but can prove that the $A_n$ are diagonalisable, I would appreciate that as an answer as well. Above I tried to look at the special case where the dimension is even, though of course the proof for all odd and even $n$ is more valuable. Even if this is not possible, for my purposes I just need an unbounded subset $Ssubseteqmathbb Z$ for which the conclusion is proven for $nin S$, so any such approach is welcome too.
Thank you in advance!
linear-algebra eigenvalues-eigenvectors determinant diagonalization
$endgroup$
3
$begingroup$
All eigenvalues distinct is a sufficient but not necessary condition for a matrix to be diagonalizable.
$endgroup$
– Henning Makholm
Apr 24 at 0:23
$begingroup$
@HenningMakholm that's a very good point. But before the responses to the question, it's the only method I knew (hence all my approaches were based on that).
$endgroup$
– YiFan
Apr 24 at 4:37
add a comment |
$begingroup$
When trying to solve a physics problem on decoupling a system of ODEs, I found myself needing to address the following problem:
Let $A_nin M_n(mathbb R)$ be the matrix with all $1$s above its main diagonal, all $-1$s below its diagonal, and $0$s everywhere else. Is $A_n$ always diagonalisable? If so, what is its diagonalisation (equivalently: what are its eigenvalues and corresponding eigenvectors)?
For example,
$$A_3=beginbmatrix0&1&0\-1&0&1\0&-1&0endbmatrix,quad A_5beginbmatrix0&1&0&0&0\-1&0&1&0&0\0&-1&0&1&0\0&0&-1&0&1\0&0&0&-1&0endbmatrix.$$
Assuming my code is correct, Mathematica has been able to verify that $A_n$ is always diagonalisable up to $n=1000$. If we use $chi_n(t)inmathbb Z[t]$ to denote the characteristic polynomial of $A_n$, a straightforward evaluation also shows that
$$chi_n(t)=-tchi_n-1(t)+chi_n-2(t)tag1$$
for all $ngeq4$. Furthermore, note that $A_n=-A_n^t$ so that, in the case where the dimension is even,
$$det(A_2n-lambda I)=det(A_2n^t-lambda I)=det(-A_2n-lambda I)=det(A_2n+lambda I).$$
This implies that whenever $lambda$ is an eigenvalue of $A_2n$, so is $-lambda$. In other words, $chi_2n(t)$ is always of the form $(t^2-lambda _1^2)(t^2-lambda_2^2)dotsm(t^2-lambda_n^2)$ for some $lambda_i$.
And this is where I am stuck. In order for $A_n$ to be diagonalisable, we must have that all the eigenvalues are distinct, but trying to use the recurrence $(1)$ and strong induction, or trying to use the formula for the even case have not helped at all. It seems like the most probable line of attack would be to somehow show that
$$chi_2n'(t)=2tsum_k=1^nfracchi_2n(t)t^2-lambda_k^2$$
never shares a common zero with $chi_2n$ (which would resolve the even case), though I don't see how to make this work.
Note: I do not have any clue how to actually find the eigenvalues/eigenvectors even in the case where the $A_n$ are diagonalisable. As such even if someone cannot answer the second part of the question, but can prove that the $A_n$ are diagonalisable, I would appreciate that as an answer as well. Above I tried to look at the special case where the dimension is even, though of course the proof for all odd and even $n$ is more valuable. Even if this is not possible, for my purposes I just need an unbounded subset $Ssubseteqmathbb Z$ for which the conclusion is proven for $nin S$, so any such approach is welcome too.
Thank you in advance!
linear-algebra eigenvalues-eigenvectors determinant diagonalization
$endgroup$
When trying to solve a physics problem on decoupling a system of ODEs, I found myself needing to address the following problem:
Let $A_nin M_n(mathbb R)$ be the matrix with all $1$s above its main diagonal, all $-1$s below its diagonal, and $0$s everywhere else. Is $A_n$ always diagonalisable? If so, what is its diagonalisation (equivalently: what are its eigenvalues and corresponding eigenvectors)?
For example,
$$A_3=beginbmatrix0&1&0\-1&0&1\0&-1&0endbmatrix,quad A_5beginbmatrix0&1&0&0&0\-1&0&1&0&0\0&-1&0&1&0\0&0&-1&0&1\0&0&0&-1&0endbmatrix.$$
Assuming my code is correct, Mathematica has been able to verify that $A_n$ is always diagonalisable up to $n=1000$. If we use $chi_n(t)inmathbb Z[t]$ to denote the characteristic polynomial of $A_n$, a straightforward evaluation also shows that
$$chi_n(t)=-tchi_n-1(t)+chi_n-2(t)tag1$$
for all $ngeq4$. Furthermore, note that $A_n=-A_n^t$ so that, in the case where the dimension is even,
$$det(A_2n-lambda I)=det(A_2n^t-lambda I)=det(-A_2n-lambda I)=det(A_2n+lambda I).$$
This implies that whenever $lambda$ is an eigenvalue of $A_2n$, so is $-lambda$. In other words, $chi_2n(t)$ is always of the form $(t^2-lambda _1^2)(t^2-lambda_2^2)dotsm(t^2-lambda_n^2)$ for some $lambda_i$.
And this is where I am stuck. In order for $A_n$ to be diagonalisable, we must have that all the eigenvalues are distinct, but trying to use the recurrence $(1)$ and strong induction, or trying to use the formula for the even case have not helped at all. It seems like the most probable line of attack would be to somehow show that
$$chi_2n'(t)=2tsum_k=1^nfracchi_2n(t)t^2-lambda_k^2$$
never shares a common zero with $chi_2n$ (which would resolve the even case), though I don't see how to make this work.
Note: I do not have any clue how to actually find the eigenvalues/eigenvectors even in the case where the $A_n$ are diagonalisable. As such even if someone cannot answer the second part of the question, but can prove that the $A_n$ are diagonalisable, I would appreciate that as an answer as well. Above I tried to look at the special case where the dimension is even, though of course the proof for all odd and even $n$ is more valuable. Even if this is not possible, for my purposes I just need an unbounded subset $Ssubseteqmathbb Z$ for which the conclusion is proven for $nin S$, so any such approach is welcome too.
Thank you in advance!
linear-algebra eigenvalues-eigenvectors determinant diagonalization
linear-algebra eigenvalues-eigenvectors determinant diagonalization
asked Apr 23 at 22:43
YiFanYiFan
5,6842831
5,6842831
3
$begingroup$
All eigenvalues distinct is a sufficient but not necessary condition for a matrix to be diagonalizable.
$endgroup$
– Henning Makholm
Apr 24 at 0:23
$begingroup$
@HenningMakholm that's a very good point. But before the responses to the question, it's the only method I knew (hence all my approaches were based on that).
$endgroup$
– YiFan
Apr 24 at 4:37
add a comment |
3
$begingroup$
All eigenvalues distinct is a sufficient but not necessary condition for a matrix to be diagonalizable.
$endgroup$
– Henning Makholm
Apr 24 at 0:23
$begingroup$
@HenningMakholm that's a very good point. But before the responses to the question, it's the only method I knew (hence all my approaches were based on that).
$endgroup$
– YiFan
Apr 24 at 4:37
3
3
$begingroup$
All eigenvalues distinct is a sufficient but not necessary condition for a matrix to be diagonalizable.
$endgroup$
– Henning Makholm
Apr 24 at 0:23
$begingroup$
All eigenvalues distinct is a sufficient but not necessary condition for a matrix to be diagonalizable.
$endgroup$
– Henning Makholm
Apr 24 at 0:23
$begingroup$
@HenningMakholm that's a very good point. But before the responses to the question, it's the only method I knew (hence all my approaches were based on that).
$endgroup$
– YiFan
Apr 24 at 4:37
$begingroup$
@HenningMakholm that's a very good point. But before the responses to the question, it's the only method I knew (hence all my approaches were based on that).
$endgroup$
– YiFan
Apr 24 at 4:37
add a comment |
3 Answers
3
active
oldest
votes
$begingroup$
The matrix $A_n$ is a tridiagonal Toeplitz matrix with diagonal entries $delta = 0$ and off-diagonal entries $tau = 1$ and $sigma = -1$. Hence, we can use the formula in this paper to show that the eigenvalues are $$lambda_k = 2icosleft(dfrackpin+1right),$$ for $k = 1,ldots,n$, and the corresponding eigenvectors $v_1,ldots,v_n$ have entries $$v_k[m] = i^msinleft(dfracmkpin+1right).$$
$endgroup$
1
$begingroup$
There's probably a connection to be made between these matrices and the Chebyshev polynomials, but I can't quite nail it down. Note that roots of the Chebyshev polynomial $T_n(lambda)$ are given by $lambda = cos( k pi/n)$ for $k in mathcalZ$. It's also possible to show that the characteristic polynomials $P_n(lambda)$ of the given matrices obey the recursion relation $P_n(lambda) = lambda P_n-1(lambda) + P_n-2(lambda)$; this is quite similar to the Chebyshev recursion relation $T_n(lambda) = 2 lambda T_n-1(lambda) - T_n-2(lambda)$.
$endgroup$
– Michael Seifert
Apr 24 at 13:58
add a comment |
$begingroup$
All those matrices are anti-symmetric and therefore they are normal matrices. And every normal matrix is diagonalizable over $mathbb C$, by the spectral theorem.
$endgroup$
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
4
$begingroup$
I would have been surprised if you had not accepted that answer, since it provides more information than mine.
$endgroup$
– José Carlos Santos
Apr 23 at 23:08
add a comment |
$begingroup$
Using that your matrices are skew symmetric, you get that these matrices are diagonalizable. See the section spectral theory on this Wikipedia article.
New contributor
$endgroup$
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
I think, compared to the above detailed answer, the only interest of mine is to show that you could have found all these informations on your own using google, as soon as you would have observed that your matrices were skew-symmetric,
$endgroup$
– gcousin
Apr 24 at 17:30
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "69"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3198918%2fare-these-square-matrices-always-diagonalisable%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
The matrix $A_n$ is a tridiagonal Toeplitz matrix with diagonal entries $delta = 0$ and off-diagonal entries $tau = 1$ and $sigma = -1$. Hence, we can use the formula in this paper to show that the eigenvalues are $$lambda_k = 2icosleft(dfrackpin+1right),$$ for $k = 1,ldots,n$, and the corresponding eigenvectors $v_1,ldots,v_n$ have entries $$v_k[m] = i^msinleft(dfracmkpin+1right).$$
$endgroup$
1
$begingroup$
There's probably a connection to be made between these matrices and the Chebyshev polynomials, but I can't quite nail it down. Note that roots of the Chebyshev polynomial $T_n(lambda)$ are given by $lambda = cos( k pi/n)$ for $k in mathcalZ$. It's also possible to show that the characteristic polynomials $P_n(lambda)$ of the given matrices obey the recursion relation $P_n(lambda) = lambda P_n-1(lambda) + P_n-2(lambda)$; this is quite similar to the Chebyshev recursion relation $T_n(lambda) = 2 lambda T_n-1(lambda) - T_n-2(lambda)$.
$endgroup$
– Michael Seifert
Apr 24 at 13:58
add a comment |
$begingroup$
The matrix $A_n$ is a tridiagonal Toeplitz matrix with diagonal entries $delta = 0$ and off-diagonal entries $tau = 1$ and $sigma = -1$. Hence, we can use the formula in this paper to show that the eigenvalues are $$lambda_k = 2icosleft(dfrackpin+1right),$$ for $k = 1,ldots,n$, and the corresponding eigenvectors $v_1,ldots,v_n$ have entries $$v_k[m] = i^msinleft(dfracmkpin+1right).$$
$endgroup$
1
$begingroup$
There's probably a connection to be made between these matrices and the Chebyshev polynomials, but I can't quite nail it down. Note that roots of the Chebyshev polynomial $T_n(lambda)$ are given by $lambda = cos( k pi/n)$ for $k in mathcalZ$. It's also possible to show that the characteristic polynomials $P_n(lambda)$ of the given matrices obey the recursion relation $P_n(lambda) = lambda P_n-1(lambda) + P_n-2(lambda)$; this is quite similar to the Chebyshev recursion relation $T_n(lambda) = 2 lambda T_n-1(lambda) - T_n-2(lambda)$.
$endgroup$
– Michael Seifert
Apr 24 at 13:58
add a comment |
$begingroup$
The matrix $A_n$ is a tridiagonal Toeplitz matrix with diagonal entries $delta = 0$ and off-diagonal entries $tau = 1$ and $sigma = -1$. Hence, we can use the formula in this paper to show that the eigenvalues are $$lambda_k = 2icosleft(dfrackpin+1right),$$ for $k = 1,ldots,n$, and the corresponding eigenvectors $v_1,ldots,v_n$ have entries $$v_k[m] = i^msinleft(dfracmkpin+1right).$$
$endgroup$
The matrix $A_n$ is a tridiagonal Toeplitz matrix with diagonal entries $delta = 0$ and off-diagonal entries $tau = 1$ and $sigma = -1$. Hence, we can use the formula in this paper to show that the eigenvalues are $$lambda_k = 2icosleft(dfrackpin+1right),$$ for $k = 1,ldots,n$, and the corresponding eigenvectors $v_1,ldots,v_n$ have entries $$v_k[m] = i^msinleft(dfracmkpin+1right).$$
edited Apr 23 at 22:58
answered Apr 23 at 22:52
JimmyK4542JimmyK4542
41.8k248110
41.8k248110
1
$begingroup$
There's probably a connection to be made between these matrices and the Chebyshev polynomials, but I can't quite nail it down. Note that roots of the Chebyshev polynomial $T_n(lambda)$ are given by $lambda = cos( k pi/n)$ for $k in mathcalZ$. It's also possible to show that the characteristic polynomials $P_n(lambda)$ of the given matrices obey the recursion relation $P_n(lambda) = lambda P_n-1(lambda) + P_n-2(lambda)$; this is quite similar to the Chebyshev recursion relation $T_n(lambda) = 2 lambda T_n-1(lambda) - T_n-2(lambda)$.
$endgroup$
– Michael Seifert
Apr 24 at 13:58
add a comment |
1
$begingroup$
There's probably a connection to be made between these matrices and the Chebyshev polynomials, but I can't quite nail it down. Note that roots of the Chebyshev polynomial $T_n(lambda)$ are given by $lambda = cos( k pi/n)$ for $k in mathcalZ$. It's also possible to show that the characteristic polynomials $P_n(lambda)$ of the given matrices obey the recursion relation $P_n(lambda) = lambda P_n-1(lambda) + P_n-2(lambda)$; this is quite similar to the Chebyshev recursion relation $T_n(lambda) = 2 lambda T_n-1(lambda) - T_n-2(lambda)$.
$endgroup$
– Michael Seifert
Apr 24 at 13:58
1
1
$begingroup$
There's probably a connection to be made between these matrices and the Chebyshev polynomials, but I can't quite nail it down. Note that roots of the Chebyshev polynomial $T_n(lambda)$ are given by $lambda = cos( k pi/n)$ for $k in mathcalZ$. It's also possible to show that the characteristic polynomials $P_n(lambda)$ of the given matrices obey the recursion relation $P_n(lambda) = lambda P_n-1(lambda) + P_n-2(lambda)$; this is quite similar to the Chebyshev recursion relation $T_n(lambda) = 2 lambda T_n-1(lambda) - T_n-2(lambda)$.
$endgroup$
– Michael Seifert
Apr 24 at 13:58
$begingroup$
There's probably a connection to be made between these matrices and the Chebyshev polynomials, but I can't quite nail it down. Note that roots of the Chebyshev polynomial $T_n(lambda)$ are given by $lambda = cos( k pi/n)$ for $k in mathcalZ$. It's also possible to show that the characteristic polynomials $P_n(lambda)$ of the given matrices obey the recursion relation $P_n(lambda) = lambda P_n-1(lambda) + P_n-2(lambda)$; this is quite similar to the Chebyshev recursion relation $T_n(lambda) = 2 lambda T_n-1(lambda) - T_n-2(lambda)$.
$endgroup$
– Michael Seifert
Apr 24 at 13:58
add a comment |
$begingroup$
All those matrices are anti-symmetric and therefore they are normal matrices. And every normal matrix is diagonalizable over $mathbb C$, by the spectral theorem.
$endgroup$
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
4
$begingroup$
I would have been surprised if you had not accepted that answer, since it provides more information than mine.
$endgroup$
– José Carlos Santos
Apr 23 at 23:08
add a comment |
$begingroup$
All those matrices are anti-symmetric and therefore they are normal matrices. And every normal matrix is diagonalizable over $mathbb C$, by the spectral theorem.
$endgroup$
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
4
$begingroup$
I would have been surprised if you had not accepted that answer, since it provides more information than mine.
$endgroup$
– José Carlos Santos
Apr 23 at 23:08
add a comment |
$begingroup$
All those matrices are anti-symmetric and therefore they are normal matrices. And every normal matrix is diagonalizable over $mathbb C$, by the spectral theorem.
$endgroup$
All those matrices are anti-symmetric and therefore they are normal matrices. And every normal matrix is diagonalizable over $mathbb C$, by the spectral theorem.
answered Apr 23 at 22:56
José Carlos SantosJosé Carlos Santos
178k24139253
178k24139253
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
4
$begingroup$
I would have been surprised if you had not accepted that answer, since it provides more information than mine.
$endgroup$
– José Carlos Santos
Apr 23 at 23:08
add a comment |
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
4
$begingroup$
I would have been surprised if you had not accepted that answer, since it provides more information than mine.
$endgroup$
– José Carlos Santos
Apr 23 at 23:08
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
4
4
$begingroup$
I would have been surprised if you had not accepted that answer, since it provides more information than mine.
$endgroup$
– José Carlos Santos
Apr 23 at 23:08
$begingroup$
I would have been surprised if you had not accepted that answer, since it provides more information than mine.
$endgroup$
– José Carlos Santos
Apr 23 at 23:08
add a comment |
$begingroup$
Using that your matrices are skew symmetric, you get that these matrices are diagonalizable. See the section spectral theory on this Wikipedia article.
New contributor
$endgroup$
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
I think, compared to the above detailed answer, the only interest of mine is to show that you could have found all these informations on your own using google, as soon as you would have observed that your matrices were skew-symmetric,
$endgroup$
– gcousin
Apr 24 at 17:30
add a comment |
$begingroup$
Using that your matrices are skew symmetric, you get that these matrices are diagonalizable. See the section spectral theory on this Wikipedia article.
New contributor
$endgroup$
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
I think, compared to the above detailed answer, the only interest of mine is to show that you could have found all these informations on your own using google, as soon as you would have observed that your matrices were skew-symmetric,
$endgroup$
– gcousin
Apr 24 at 17:30
add a comment |
$begingroup$
Using that your matrices are skew symmetric, you get that these matrices are diagonalizable. See the section spectral theory on this Wikipedia article.
New contributor
$endgroup$
Using that your matrices are skew symmetric, you get that these matrices are diagonalizable. See the section spectral theory on this Wikipedia article.
New contributor
New contributor
answered Apr 23 at 22:57
gcousingcousin
1412
1412
New contributor
New contributor
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
I think, compared to the above detailed answer, the only interest of mine is to show that you could have found all these informations on your own using google, as soon as you would have observed that your matrices were skew-symmetric,
$endgroup$
– gcousin
Apr 24 at 17:30
add a comment |
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
I think, compared to the above detailed answer, the only interest of mine is to show that you could have found all these informations on your own using google, as soon as you would have observed that your matrices were skew-symmetric,
$endgroup$
– gcousin
Apr 24 at 17:30
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
Thank you very much for the quick response! I hope you don't mind that I accepted JimmyK4542's answer, since it also gives explicitly the eigenvectors and eigenvalues.
$endgroup$
– YiFan
Apr 23 at 23:06
$begingroup$
I think, compared to the above detailed answer, the only interest of mine is to show that you could have found all these informations on your own using google, as soon as you would have observed that your matrices were skew-symmetric,
$endgroup$
– gcousin
Apr 24 at 17:30
$begingroup$
I think, compared to the above detailed answer, the only interest of mine is to show that you could have found all these informations on your own using google, as soon as you would have observed that your matrices were skew-symmetric,
$endgroup$
– gcousin
Apr 24 at 17:30
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3198918%2fare-these-square-matrices-always-diagonalisable%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
3
$begingroup$
All eigenvalues distinct is a sufficient but not necessary condition for a matrix to be diagonalizable.
$endgroup$
– Henning Makholm
Apr 24 at 0:23
$begingroup$
@HenningMakholm that's a very good point. But before the responses to the question, it's the only method I knew (hence all my approaches were based on that).
$endgroup$
– YiFan
Apr 24 at 4:37