Functions that are the object of discussion in calculus are mathematical objects that map a scalar number to another number
x⟶f(x). While working with operators in our quantum mechanics course, we came across terms which seemed to be polynomial functions of the position X^ and momentum P^ operators such as
or
for example in the Hamiltonian operator of the harmonic oscillator
H^=2mP^2+21mω2X^2. Polynomial terms like these are well-defined, but what about more general functions of operators? For example, can we assign any meaning to the exponential of an operator
The answer is yes! But in order to better understand the concept of functions of operators, let us consider the simpler case of matrices in finite dimensional vector spaces. In this problem, we will consider a 2-dimensional vector space, therefore the operators can be represented using 2×2 matrices. Many of the results can be generalized to infinite dimensional vector spaces but we do not concern ourselves with them now.
Inspired by the power series of the usual exponential function
ex=1+x+2!x2+3!x3+⋯=n=0∑∞n!xn, we can define the exponential of an operator as a power series
eA^=I^+A^+2!A^2+3!A^3+⋯=n=0∑∞n!A^n. As we will see in this problem, the above definition leads to plausible results that are consistent with the properties we expect and therefore is a valid definition of the exponential of an operator.
Assume we are given a diagonal matrix M^ as
M^=(λ100λ2), show that the exponential of M^ as defined in Eq. 7 is simply another diagonal matrix where its diagonal elements are the exponential of the elements of the original matrix M^. That is
eM^=(eλ100eλ2). Hint: Try to explicitly calculate a few first terms 2!M^2 and 3!M^3 in the power series of the exponential of M^ and see if you can recognize a pattern in finite summations like
n=0∑1n!M^nn=0∑2n!M^nn=0∑3n!M^n=I^+M^,=I^+M^+2!M^2,=I^+M^+2!M^2+3!M^3. Solution
It is straightforward to see that
M^n=(λ1n00λ2n), by direct calculation.
This means we have
I^+M^I^+M^+2!M^2I^+M^+2!M^2+3!M^3=(1+λ1001+λ2),=(1+λ1+2!λ12001+λ2+2!λ22),=(1+λ1+2!λ12+3!λ13001+λ2+2!λ22+3!λ23), From this, we can see that the summation in Eq. 7 always results in a diagonal matrix and we have
n=0∑∞n!M^n=(∑n=0∞n!λ1n00∑n=0∞n!λ2n)=(eλ100eλ2)
After working out the example using the exponential function, we can see that the definition of an arbitrary function of an operator can as well be defined for any function f which has a power series representation
f(x)=n=0∑∞cnxn, as
f(A^)=n=0∑∞cnA^n. Using the general definitions in Eq. 14 and Eq. 15, argue that the function of a diagonal matrix M^ is simply a diagonal matrix with the corresponding function applying to the diagonal elements. That is
f(M^)=(f(λ1)00f(λ2)). Note that from this point on we use the same letter f as the notation for both a function of an operator f(A^) and a function of a scalar f(x). These two are really different things and they should not be confused with each other. The first one maps an operator to an operator and the second one maps a scalar to a scalar. However, we continue using the letter f for both because it is clear from the context what we mean.
Solution
It is easy to see this from the previous result. For the diagonal matrix M^, we have
M^n=(λ1n00λ2n). Therefore, the summation
f(M^)=n=0∑∞cnM^n will become
f(M^)=n=0∑∞cnM^n=(∑n=0∞cnλ1n00∑n=0∞cnλ2n)=(f(λ1)00f(λ2)).
The results we have proved so far apply to diagonal matrices. But what about other matrices? Can Eq. 15 be meaningful for operators A^ which are not necessarily diagonal? For the upcoming tasks, consider a more general matrix A^ which is not necessarily diagonal but can be diagonalized using an invertible matrix P^ as
A^=P^D^P^−1. One of the problems in Problem Set 5 was about diagonalizing a matrix in this fashion and in the following, you will see a few reasons for our interest in such matrices. Not all matrices are diagonalizable but almost all operators we work with in quantum mechanics are either diagonalizable or composed out of diagonalizable operators.
Show that for a diagonalizable matrix A^=P^D^P^−1, if we use the definition of a function of an operator in Eq. 15, we have
f(A^)=P^f(D^)P^−1 which means given the diagonalized form of A^, if we want to calculate f(A^) we only need to apply f to the diagonal part in the diagonalized representation of A^. This is a powerful result for calculating a function of any diagonalizable operator.
Hint: This is not as difficult as it seems. The only thing you have to do is to again try to calculate a few terms in the summation in Eq. 15. Use the diagonalized form of A^ and try to evaluate terms like A^2=A^A^ and A^3=A^A^A^. The operator P^ and its inverse will help simplifying these terms. Can you find a simple representation of A^n in terms of P^, P^−1, and powers of D^?
Solution
Let us have a look at some of the terms! The fact that P is invertible helps us simplifying the terms
A^2=A^A^=P^D^I^P^−1P^D^P^−1=P^D^D^P^−1=P^D^2P^−1, and similarly
A^3=A^A^A^=P^D^I^P^−1P^D^I^P^−1P^D^P^−1=P^D^D^D^P^−1=P^D^3P^−1. Therefore, we have the general relation
A^n=P^D^nP^−1 because the P^ and P^−1 operators in the middle always cancel out.
This means we can write
f(A^)=n=0∑∞cnA^n=n=0∑∞cnP^D^nP^−1=P^[n=0∑∞cnD^n]P^−1=P^[f(D^)]P^−1 Use the result of the previous task to calculate
where
A=(4213) is the matrix from the Problem Set 5. And the diagonalized form of it was
A=PDP−1=(111−2)(5002)(2/31/31/3−1/3). Solution
we have
eA=PeDP−1=(111−2)(e500e2)(2/31/31/3−1/3)=31(2e5+e22e5−2e2e5−e2e5+2e2)
You can see that calculating eA directly using the Taylor series will be much more difficult compared to calculating it using the diagonal form of A.
We will now end this problem by proving a very important result in functions of operators which is about the action of them on eigenvectors.
Show that if ∣v⟩ is a eigenvector of A^ with the eigenvalue λ, meaning that
A^∣v⟩=λ∣v⟩, then ∣v⟩ is also an eigenvector of any function of that operator f(A^) with the eigenvalue f(λ). That is
f(A^)∣v⟩=f(λ)∣v⟩. Hint: Apply the operator f(A^) to the eigenvector ∣v⟩ and use the definition of the function of an operator in Eq. 15. Because we have A^∣v⟩=λ∣v⟩, we can easily evaluate terms like A^n∣v⟩ in the power series expansion. Then check if you see any similarities between the value you calculate in the power series and the corresponding definition of the function on normal scalars in Eq. 14.
Solution
The action of polynomials operator forms A^n on the eigenvector is trivial.
A^∣v⟩A^2∣v⟩A^3∣v⟩=λ∣v⟩=A^A^∣v⟩=λA^∣v⟩=λ2∣v⟩=A^A^A^∣v⟩=λA^A^∣v⟩=λ2A^∣v⟩=λ3∣v⟩ Therefore,
A^n∣v⟩=λn∣v⟩ This means
f(A^)∣v⟩=n=0∑∞cnA^n∣v⟩=n=0∑∞cnλn∣v⟩=[n=0∑∞cnλn]∣v⟩=f(λ)∣v⟩ Verify the result of the previous task by applying the operator eA you calculated in part (d) to the eigenvectors of A (You can go back to the solutions of the Problem Set 5 to check out the eigenvectors and the corresponding eigenvalues of A). This means you have to show
eA(11)=e5(11) and
eA(1−2)=e2(1−2). Solution
We need to manually calculate the matrix products. We have
eA(11)=31(2e5+e22e5−2e2e5−e2e5+2e2)(11)=31(3e53e5)=e5(11), and
eA(1−2)=31(2e5+e22e5−2e2e5−e2e5+2e2)(1−2)=31(3e2−6e2)=e2(1−2)
This is extremely powerful! It shows that eigenvectors and eigenvalues are even more useful than we previously thought. Once you find the eigenvectors and the eigenvalues of an operator, you can evaluate the action of any function of that operator on the eigenvectors without the need to explicitly calculating f(A^) and by simply substituting f(A^) with f(λ) as in Eq.31.