Academia.eduAcademia.edu

Incorporation of New Information in an Approximate Hamiltonian

Additional information about the eigenvalues and eigenvectors of a physical system demands extension of the effective Hamiltonian in use. In this work we extend the effective Hamiltonian that describes partially a physical system so that the new Hamiltonian comprises, in addition to the information in the old Hamiltonian, new information, available by means of experiment or theory. A simple expression of the enlarged Hamiltonian, which does not involve matrix inversion, is obtained. It is also shown that the Lee-Suzuki transformation effectively put the initial Hamiltonian in a diagonal block form.

Incorporation of New Information into an Approximate Hamiltonian C. P. Viazminsky Department of Physics University of Aleppo, Syria Sohail Baza Department of Physics University of Aleppo, Syria Abstract Additional information about the eigenvalues and eigenvectors of a physical system demands extension of the effective Hamiltonian in use. In this work we extend the effective Hamiltonian that describes partially a physical system so that the new Hamiltonian comprises, in addition to the information in the old Hamiltonian, new information, available by means of experiment or theory. A simple expression of the enlarged Hamiltonian, which does not involve matrix inversion, is obtained. It is also shown that the Lee-Suzuki transformation effectively put the initial Hamiltonian in a diagonal block form. 1. Introduction It is well known that Schrodinger equation cannot be solved analytically except in limited simple cases. This calls upon approximation methods, of which perturbation method is most common [1,2], to find approximate solutions of this equation. The perturbation method hinges on ignoring a part of the Hamiltonian H, called the perturbed part, so that the resulting equation is solvable analytically. The full Hamiltonian H is described as perturbed, whereas the simplified one is described as unperturbed. The unperturbed equation is solved analytically, and the corrections that take into account the ignored part are calculated. The set of eigenvectors {ei }1∞ of the unperturbed Hamiltonian is taken as a basis in the infinite dimensional Hilbert space of wave functions, denoted by Η ∞ . Observables pertaining to the system are represented by Hermitian matrices in terms of this basis. Unless these matrices are given by recurrence formulae, we have to be contented with finite matrix approximation, which implies truncating the infinite basis at some sufficiently large term N [12]. The space generated by the truncated basis {e µ }1N -denoted by Η and called the truncated space- will hopefully contain good approximation of all states of interest to the problem we consider. It must be noted that, whenever the eigenvalue problem is to be solved numerically, which is usually the case in physically interesting problems, truncation is inevitable task. We also note that truncating an infinite basis by a finite one with a sufficiently large number of basis elements is justified by the fact that the sequence (e N ) tends weakly to zero as N tends to infinity. This means that for every wave function ψ ∈ Η ∞ the sequence of numbers (< e N | ψ >) tends to zero as N tends to 1 infinity [3,4]. Alternatively, an upper cutoff, N, can be safely applied without seriously changing the low-lying properties [12]. Supposing that the truncated space Η is good enough to replace the full Hilbert space, the full Hamiltonian can be represented by a ( N × N ) matrix that acts in the truncated space. It must be emphasized that truncation is dictated by the computation process and is independent of the approximation method employed. It is quite likely that the approximation method may yield good estimates {E1 ,..., E d } of the first d levels, and that the accuracy decreases as the number n of the level increases, so that it becomes unacceptable for n > d (n is still much smaller than N). Assuming that some experimental facts or theoretical considerations assures that the numbers {E d +1 ,...., E n } are good estimates of the next (n-d) levels, then the question arises as how to construct a new approximate Hamiltonian whose spectrum is {E1 ,..., E n }. A second approach in approximation, namely effective Hamiltonian methods [5-12], is common in nuclear and molecular physics. This approach is based on constructing a Hamiltonian H 11 in a d-dimensional subspace of the truncated space Η that possesses d eigenvalues of the full Hamiltonian. The question also arises how to incorporate in this Hamiltonian additional data based on experiment or theory. Our goal in this work is to study the problem associated with this question in an abstract sense, and give a rigorous answer of the question we just posed. 2. Algebraic Considerations Let Π be a d-dimensional Hilbert space. Π is then isomorphic to C n . The natural embedding ( x1 ,...., x d ) t ∈ Π → ψ = ( x1 ,...., x d ,0,0,.....,0) t ∈ Η n of the space Π in the Hilbert space Η n (n > d ) is an isomorphism between Π and a d-dimensional subspace of the Hilbert space Η n . Through this isomorphism we may α  identify every vector α i ∈ Π with a vector ψ i ≡  i  ∈ Η n . Let L11 be a linear 0 operator in the Hilbert space Π. If {eµ }nµ =1 is a basis in Η n , then the operator L11 , which is represented in the basis {ei }1d by a (d × d ) matrix L11 , may also be viewed as an operator in Η n with the corresponding matrix representation L l =  11 0 Thus we have 0  . 0 n −d  α   L α  l  i  =  11 i  0  0  ∀α i ∈ Π (1) (2) Assume now that L11 possesses the discrete spectrum sp ( L11 ) = {E1 ,...., E d } to which the linearly independent eigenvectors {α 1 ,..., α d } belong. And let {E d +1 ,..., E n } be a set of numbers {ψ d +1 ,...,ψ n }, be a set of n-vectors such that the set of n-vectors 2 α  α  Ψ = {ψ 1 =  1 ,....,ψ d =  d ,ψ d +1 ,...,ψ n }  0  0 is linearly independent. The problem which we consider is how to construct an operator L12  L L =  11 (3)   L21 L22  in Η n such that the spectrum of L is {E1 ,..., E d ,...., E n }, and the corresponding set of eigenvectors is Ψ = {ψ µ }1n . We shall let the indices i, j and µ run from 1 to d, from d + 1 to n, and from 1 to n respectively. Now the eigenequation Lψ µ = E µψ µ ( µ = 1,..., n) (4) is conveniently expressed in matrix form as  L11   L21 L12 α  L22  0 a  α = b   0 a  E 0   , b  0 e  (5) where α = [α 1 | .... | α d ] is an d × d matrix whose columns are the vectors α i (i = 1,..., d ). a  b  = [ψ d +1 | .... | ψ n ] is an n × (n − d ) matrix whose columns are the vectors   ψ j ( j = d + 1,..., n). This matrix is partitioned into an d × d matrix a and an aj  (n − d ) × d matrix b, so that E ψ j =  , ( j = d + 1,..., n). and e are (d × d ) and bj  (n − d ) × (n − d ) diagonal matrices with diagonal elements Ei and E j respectively. Now the eigenequation (5) is equivalent to tow sets of equations I an II (i ) L11α = αE , (ii ) L21α = 0 (I ) (i ) L11 a + L12 b = ae , (ii ) L21 a + L22 b = be ( II ) Since the set of vectors {α 1 ,...., α d } is linearly independent, the second equation in (I), expanded as L12α i = 0 (i = 1,..., d ), shows that L21 β = 0 (∀β ∈ H d ), and hence L21 = 0. On substituting this result in the second set II, we obtain (i ) L11 a + L12 b = ae, (ii ) L22 b = be ( II )′ Now α a   = det α . det b ≠ 0, (6) det  0 b because the set of columns of this matrix, namely {ψ µ }1n , is linearly independent. Since the set of vectors {α i }1d is also linearly independent, det α ≠ 0. Hence det b ≠ 0, and b is invertible. In particular, the latter result affirms that no vector b j is zero, and that all vectors b j are eigenvectors of the operator L22 belonging to the 3 0 eigenvalues E j . However, a vector   is not an eigenvector of L unless L12 = 0, bj  as it is implied by equation ( II )′. Multiplying both equations in ( II )′ from left by b −1 we obtain L22 = beb −1 , L12 b = (ae − L11 a ). (7) Hence L (ae − L11 a )b −1  . (8) L =  11 −1  beb 0   It is noted that the matrix which we seek, L, could have been directly obtained by the equation −1 α a  E 0 α a   ,   L =  (9)  0 b  0 e  0 b  However this, in practice, is much more cumbersome than formula (8) which requires inversion of an (n − d ) × (n − d ) matrix, instead of inverting an (n × n) matrix as required by (9). In actual physical problems n-d is much smaller than d. We shall see in the next section that, when the operator L is Hermitian, a great deal of simplification could be achieved, and that the expression obtained does not involve matrix inversion at all. 3. Conditions of Diagonal Block Form We discuss here the conditions imposed on the given eigenvectors so that the matrix L, which we have constructed, is immediately in a diagonal block form: ~ L L =  11 0 0  . beb −1  (10) For this to be the case we should have in (8) L12 = 0, which is equivalent to L11 a = ae, since b is non-singular. This also could be seen from the first equation in (7), which shows that the operator L12 : Η ⊥d → Η d transforms each vector b j to a vector ( L11b) j = (ae − L11 a ) j = E j a j − L11 a j . An equation L11 a j − E j a j = 0 holds if either of the following conditions hold: 0 (i) a j = 0 or equivalentlyψ j =  . bj  (ii) The aj  eigenvector ψ j =   with a j ≠ 0 belongs bj  to an eigenvalue E j = Ei ∈ {E1 ,..., E d }. In this case a j is an eigenvector of L11 belonging also to the 4 0 degenerate eigenvalue E j , and henceψ j − ψ i =   is also an eigenvector of L bj  belonging to the degenerate eigenvalue E j = Ei . We deduce accordingly that if for some j ∈ [d + 1, n], it is given that a j ≠ 0 and that a j is not an eigenvector of L11 then L cannot immediately assume a diagonal block form. In particular if the sets {Ei }1d and {E j }nd +1 do not intersect, then all a j have to vanish, if L is to assume immediately a diagonal block form. On the other hand and if either of the above conditions (i) or (ii) hold, we may replace every vector ψ j given in the data by its projection on Π ⊥ . In all cases the block form (10) is equivalent, through a similarity transformation, to the diagonal block form (8). In fact it is easy to show that the similarity transformation effected by the matrix  1 − ab −1   Τ =  1  0 has the following properties 1. It leaves all the vectorsψ i unaltered, i.e. ψ~i ≡ Τψ i = ψ i . 0 2. It projects every vectorψ j on the subspace Π ⊥ , i.e. ψ~ j ≡ Τψ j =  . b ~ ~ 3. It transforms L to L , i.e. L = ΤLΤ −1 . 4. Incorporation of New Information into the Hamiltonian We mean by this title that given a Hermitian operator H 11 (for example the Hamiltonian of some physical system) as well as a set of numbers and a corresponding set of vectors, then our goal is to construct a new Hermitian operator H whose spectrum is composed of the spectrum of H 11 as well as these given numbers, and whose corresponding eigenvectors consist of the set of eigenvectors of H 11 in addition to the given vectors. There are, of course, some conditions which are to be satisfied by the given vectors. These conditions will be spelled out in the sequel. To state the problem concretely, let H 11 be a Hermitian operator in a Hilbert space Π, and assume that H 11 possesses the discrete spectrum Ei (i = 1,..., d ) with the corresponding linearly independent eigenvectors ψ i (i = 1,..., d ), so that H 11ψ i = Eiψ i (i = 1,..., d ). Suppose also that we have a set of numbers E j ( j = d + 1,..., n) and a set of orthogonal linearly independent vectors ψ j ( j = d + 1,...., n). We seek to construct a Hermitian operator H in Η n such that Hψ µ = E µψ µ ( µ = 1,..., n) 5 (11) Since the eigenvectors of a Hermitian operator that belong to different eigenvalues have to be orthogonal, and that which belong to the same eigenvalue can be orthogonalized by Gram-Schmidt procedure [4], we assume that the given vectors together with the eigenvectors of H 11 are orthogonal: (12) ( µ ≠ ν ; µ ,ν = 1,...., n) < ψ µ | ψ ν >= 0 For notational convenience we shall set ψ µ ≡| µ > and adopt Dirac’s notations, so that the last equation takes the form < µ | ν >= 0 ( µ ≠ ν ). To gain insight into the method of solution, we consider as a first step the operator L = H 11 + < j | j > −1 {E j | j >< j | − H 11 | j >< j |} (13) The operator | j >< j | with j fixed on one value in the range {d + 1,..., n} is the projection operator on a one-dimensional subspace of Η d +1 generated by the vector ψ j ≡| j > . Evidently, all operators appearing in equation (13) act in the Hilbert space 0 H H d +1 . In particular H 11 is identified with the operator  11  in Η d +1 . In notations  0 0 more familiar to mathematicians, equation (13) is written thus L = H 11 + ψ j −2 {E j Pj − H 11 Pj } where Pj is the projection operator on ψ j . We shall show now that the spectrum of the operator L is {E1 ,..., E d , E j } and that the corresponding eigenvectors are {| 1 >,.... | d >, | j >}. Indeed L | i > = H 11 | i > + < j | j > −1 ( E j − H 11 ) | j >< j | i > = E i | i > + < j | j > −1 [( E j − H 11 ) | j >] < j | i > = Ei | i > . L | j >= H 11 | j > + < j | j > −1 ( E j − H 11 ) | j >< j | j > = H 11 | j > + E j | j > − H 11 | j >= E j | j > . The generalization of the expression (13) to obtain an operator H in Η n that fulfill our requirements is straightforward. It can be checked easily that the following operator H = H 11 ( I n − n ∑ | j >< j |) + j = d +1 n ∑E j =1+ d j | j >< j | satisfies the eigenequation H | µ >= E µ | µ >, ( µ = 1,..., n). Indeed 6 (14) n n ( µ ≤ d ) ⇒ H | µ >= H 11 (| µ > − ∑ | j >< j | µ >) + i = d +1 ∑E j = d +1 j | j >< j | µ > = H 11 | µ >= E µ | µ > = H 11 (| µ > − n n j = d +1 j = d +1 ∑ | j >< j | µ >) + ∑ E ( µ > d ) ⇒ H | µ >= H 11 (| µ > − n n ∑| j > δ µ ) + ∑ E j j = d +1 j = d +1 j j | j >< j | µ > | j > δ jµ = H 11 (| µ > − | µ >) + E µ | µ > = Eµ | µ > . It remains thus to prove that the operator H ,which we have just constructed, is Hermitian. 5. On The Hermicity of H. Denote by Q and P respectively the projection operators from Η n onto Π and Π ⊥ , where Π ⊥ is the orthogonal complement of Π, then [4] Q + P = I n , QP = 0, Q 2 = Q, P 2 = P (15) In either basis {e µ }1n , or {ψ µ }1n ≡ {| µ >}1n in Η n the projection operator Q is represented by the matrix 0  . 0 n −d  I Q= d 0 Indeed (16) Qµν =< µ | Q | ν >= 0 if µ > d or ν > d =< i | Q | j > if µ = i ≤ d and ν = j ≤ d = δ ij In fact the representation (16) is always valid as long as every basis element does not have components in both Π and Π ⊥ . This is evident since the restriction of Q to Π is merely the identity operator. Similarly, the matrix of the projection operator P is 0  0 (17) P= d  0 I n −d   The equation P= n ∑ | j >< (18) j| j = d +1 is valid since both of its sides are projection operators on Π ⊥ . The operator H can therefore be expressed thus: H = H 11 ( I n − P ) + 7 n ∑ E j Pj = H 11Q + j = d +1 n ∑ E j Pj = QH 11Q + j = d +1 n ∑E j = d +1 j Pj . Recalling that projection operators are Hermitian, we deduce that H = H + , i.e. the operator H is equal to its adjoint operator H + , and hence is Hermitian. References [1] Landau L D and Lifshitz E M, 1975 – Quantum Mechanics. Pergamon Press, Oxford . [2] Messiah A, 1970- Quantum Mechanics. North Holland Publishing Company, Amsterdam. [3] Prugovecki E, 1971- Quantum Mechanics in Hilbert Spaces, Academic Press. [4] Fano G, 1967 - Foundation of Quantum Mechanics, McGraw Hill. [5] Lee S Y and Suzuki K, Physics Letters B91 (1980). [6] Suzuki K and Lee S Y, Progress of Theoretical Physics Vol. 64 (1980). [7] Suzuki K, Progress of Theoretical Physics Vol. 68 (1982). [8] Suzuki K and Okamoto R, Progress of Theoretical Physics Vol. 71 (1984). [9] Suzuki K, Okamoto R and Kumagai H, Journal of Nuclear Physics A580 (1994). [10] Fields T J, Vary J P and Gupta K S, Modern Physics Letters A11 (1996). [11] Navratil P and Barrett B R, Physics Letters B369 (1996). [12] Viazminsky C and Vary J P, Journal of Mathematical Physics, Vol. 41 (2000). 8