TransWikia.com

Help understanding proof in simultaneous diagonalization

Physics Asked on June 26, 2021

The proof is from Principles of Quantum Mechanics by Shankar. The theorem is:

If $Omega$ and $Lambda$ are two commuting Hermitian operators, there exists (at least) a basis of common eigenvectors that diagonalizes them both.

The proof is:
Consider first the case where at least one of the operators is nondegenerate, i.e. to a given eigenvalue, there is just one eigenvector, up to a scale. Let us assume $Omega$ is nondegenerate. Consider any one of its eigenvectors:

$$Omegaleft|omega_irightrangle=omega_ileft|omega_irightrangle$$
$$LambdaOmegaleft|omega_irightrangle=omega_iLambdaleft|omega_irightrangle$$
Since $[Omega,Lambda]=0$
$$OmegaLambdaleft|omega_irightrangle=omega_iLambdaleft|omega_irightrangle$$

i.e., $Lambdaleft|omega_irightrangle$ is an eigenvector with eigenvalue $omega_i$. Since this vector is unique up to a scale,

$$Lambdaleft|omega_irightrangle=lambda_ileft|omega_irightrangle$$

Thus $left|omega_irightrangle$ is also an eigenvector of $Lambda$ with eigenvalue $lambda_i$…

What I do not understand is the statement/argument “Since this vector is unique up to a scale.” I do not see how the argument allows to state the equation following it. What axiom or what other theorem is he using when he states “since this vector is unique up to a scale”?

4 Answers

When $lambda_1$ is an eigenvalue of a matrix and $v_1$ and $v_2$ are the components of the corresponding eigenvector, then the following equation holds:

$begin{pmatrix} a-lambda_1 & b c &d-lambda_1 end{pmatrix}begin{pmatrix} v_1 v_2 end{pmatrix}=begin{pmatrix} 0 0 end{pmatrix}$

Now when you scale up the eigenvector (say by three) it looks like this:

$begin{pmatrix} a-lambda_1 & b c &d-lambda_1 end{pmatrix}begin{pmatrix} 3v_1 3v_2 end{pmatrix}=begin{pmatrix} 0 0 end{pmatrix}$

This you can write as

$3 begin{pmatrix} a-lambda_1 & b c &d-lambda_1 end{pmatrix}begin{pmatrix} v_1 v_2 end{pmatrix}=begin{pmatrix} 0 0 end{pmatrix}$

But the matrix multiplied with the eigenvector still yields the zero vector!

This is what he meant when he said "Since this vector is unique up to a scale.": any scaled up eigenvector of a matrix is still an eigenvector. And how does the last equation follow from it?

When you write $$OmegaLambdaleft|omega_irightrangle=omega_iLambdaleft|omega_irightrangle$$ then you know that $Lambdaleft|omega_irightrangle$ gives you a vector which is an eigenvector of $Omega$. But you said that $Omega$ is nondegenerate, so for any $omega_i$ there is a unique $left|omega_irightrangle$. What this means is that this eigenvector you get by applying $Lambdaleft|omega_irightrangle$ must be $left|omega_irightrangle$. Luckily, any eigenvector which is scaled up (here by $lambda_i$) is still an eigenvector, so that you get

$$Omegalambda_ileft|omega_irightrangle=omega_i lambda_i left|omega_irightrangle$$

or

$$Lambdaleft|omega_irightrangle=lambda_ileft|omega_irightrangle $$

Correct answer by user42076 on June 26, 2021

Note that he explains above: "Consider first the case where at least one of the operators is nondegenerate, i.e. to a given eigenvalue, there is just one eigenvector, up to a scale." So he uses the assumption that the operator is nondegenerate and the definition of nondegeneracy (or a statement equivalent to the definition of nondegeneracy, if you use a different definition).

The phrase "to a given eigenvalue, there is just one eigenvector, up to a scale" means that any two eigenvectors with the same eigenvalue $omega_i$ coincide up to a factor. Then $lambda_i$ is such a factor.

Answered by akhmeteli on June 26, 2021

Since the vector $Lambda | omega_i rangle $ has the same eigenvalue as $| omega_i rangle $, it must be in the same invariant subspace as $| omega_i rangle $, which Shankar takes to be one dimensional.

Answered by ZachMcDargh on June 26, 2021

I don't know exsactly what it means with scalar but remember that:

A vector $|Psirangle$ is invariant up to a phase because a global phase is always ruled out when you calculate, for example, with the state $e^{itheta}|Psirangle$, the expected value of and observable $A$ using this state is

$$langle A rangle = langlePsi | e^{-itheta} A e^{itheta} |Psi rangle = langlePsi | A | Psirangle,$$

which is the same result as if you calculate the expected value by just using the vector $|Psirangle$. So for all practical purposes, you can use just a state $|Psirangle$ in your punch line of your proof.

Answered by user2820579 on June 26, 2021

Add your own answers!

Ask a Question

Get help from others!

© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP