Line 123: Line 123:
 
It's like choosing which coordinates to label
 
It's like choosing which coordinates to label
 
x_1, x_2, x_3, etc.  It is an arbitrary choice.
 
x_1, x_2, x_3, etc.  It is an arbitrary choice.
 +
 +
Question:
 +
 +
How can I find the roots of the characteristic
 +
polynomial for p. 348, #13?  It is a mess.
 +
 +
Answer:
 +
 +
I ran the commands
 +
 +
<PRE>
 +
with(linalg):
 +
a := matrix([[14-r,4,-2],[4,14-r,2],[-2,2,17-r]]);
 +
det(a);
 +
solve( det(a)=0, r);
 +
</PRE>
 +
 +
and found the roots to be 18,18,9.  After that, I noticed
 +
that if I multiply the second row of  (A - rI)  to the first
 +
row, I get the top row in the determinant to be
 +
 +
<PRE>
 +
18-r  18-r  0
 +
</PRE>
 +
 +
so that if I expand along that row, I get an  18-r  factored out nicely
 +
and the rest is easy.
  
 
[[2010 MA 527 Bell|Back to the MA 527 start page]]  
 
[[2010 MA 527 Bell|Back to the MA 527 start page]]  

Revision as of 08:53, 15 September 2010

Homework 3 collaboration area

Question from student regarding HW #3:

When the directions state to find the spectrum, is it asking for the spectrum of the original matrix or the spectrum of the symmetric, skew-symmetric, orthoganal, hermitian, skew-hermitian, or unitary matrix?

Answer from Bell:

You'll find the definition of the spectrum on page 334. It is just another word for the set of all eigenvalues. The problem is asking you to find the spectrum for the original matrix. Then determine the type of the matrix and verify if the eigenvalues satisfy the conditions of Theorems 1 and 5.

Question from student regarding HW#3:

How much detail is sufficient on page 339, problem 30?

Answer from Bell:

You'll need to explain in words, using results from the book as evidence to back up your statements. It will take three or four sentences to explain and your explanation should convince anyone who has read the book that you are correct. If somebody else in the class reads your explanation, they should say, oh yes! That makes perfect sense.

Question from student:

I am stuck on where to start with HW3, lesson 7, #30, the proof that the inverse of a square matrix exists iff no eigenvalues are zero.

Answer from Bell:

Start with the definition of an eigenvalue. If r is an eigenvalue, then

det(A-rI)=0.

If r=0, what does that equation say about the matrix A?

Question:

Is this an alternative approach?: Assume A^-1 exist and multiply both sides of equation (3) from p. 335 by A^-1 to get A^-1 * [(A-rI)x] = A^-1 * 0. Then ask what is A^-1? I got a solution that says A^-1 = (1/r)*I but this cannot be right.....can it? --Rayala 18:32, 13 September 2010 (UTC)

Answer:

In order to say that A^(-1) exists, you'll need to know that the determinant of A is non-zero. Once you establish that, when you multiply as you suggest, be sure to distribute through the sum properly and use the rules of scalar and matrix multiplication. When I do it, I get

$ A^{-1}(A-rI)x=\left(A^{-1}A-A^{-1}(rI)\right)x=(I - r A^{-1})x=0. $

So

$ rA^{-1}x = Ix=x. $

Now divide by r.

Question from student:

Problem 11 on p 348 is asking if the matrix is symmetric, skew-symmetric, or orthogonal. I found the eigenvalues to be a repeated real root, but the matrix is neither of the 3.

Answer from Bell:

That is correct. The book makes the point that if the matrix does not satisfy the conditions of the theorems, then the theorems do not apply. The eigenvalues may or may not satisfy any special conditions. You have done what the problem asks for if you add the clause, "Theorems 1 and 5 do not apply."

Question from student:

Problem 12 has some really nasty eigenvalues. Am I doing something wrong?

Answer from Bell:

You'll be surprised how nice it all turns out. Be sure to use the trig identity


$ \cos^2\theta+\sin^2\theta=1. $

You'll need to use it once to simplify a cos^2 + sin^2 term, and again to clean up the square root of cos^2-1. (You'll get complex eigenvalues that way.)

Question:

Page 348, problem 11: the book has a part of the answer as "defect 1", what does this mean?

Answer:

You'll find the definition of "defect" at the bottom of page 337. It is the algebraic multiplicity minus the geometric multiplicity. A matrix with any positive defects is called defective. (Really.)

Question:

I'm having trouble finding the eigenvalues for the 4x4 matrix of p. 338: 23. Any tips?

Answer:

Notice that the matrix is lower triangular. Hence, when you subtract lambda down the diagonal and take the determinant, you get the product of the diagonal entries -- all factored and ready to go. The roots are just the diagonal entries of the matrix.

Question:

I have a question regarding diagonalization of a matrix. The X matrix is determined from the eigenvectors, but how do you know which order to use them? I've noticed that the diagonal matrix is different depending on which order I choose. Can a matrix have more than one "diagonalized" matrix? Or are these different matrices I calculate somehow related?

Answer:

You can put those eigenvectors in any order (and if you don't normalize them, they can be multiplied by non-zero constants, too). Hence, X is not uniquely determined. It must, however, be a matrix whose columns are a linearly independent set of n eigenvectors for the matrix. (If an nxn matrix does not have a full set of n independent eigenvectors, it is not diagonalizable.)

The diagonalized matrix is unique up to the order the eigenvalues appear down the diagonal.

It's like choosing which coordinates to label x_1, x_2, x_3, etc. It is an arbitrary choice.

Question:

How can I find the roots of the characteristic polynomial for p. 348, #13? It is a mess.

Answer:

I ran the commands

with(linalg):
a := matrix([[14-r,4,-2],[4,14-r,2],[-2,2,17-r]]);
det(a);
solve( det(a)=0, r);

and found the roots to be 18,18,9. After that, I noticed that if I multiply the second row of (A - rI) to the first row, I get the top row in the determinant to be

18-r   18-r   0

so that if I expand along that row, I get an 18-r factored out nicely and the rest is easy.

Back to the MA 527 start page

To Rhea Course List

Alumni Liaison

Sees the importance of signal filtering in medical imaging

Dhruv Lamba, BSEE2010