In the next chapter we will formulate the general principles of quantum mechanics. Here, we introduce the necessary mathematical framework, which concerns vectors, scalar products (a complex version of the dot product), and operators. We describe these entities in the powerful Dirac notation, which is widely used throughout quantum mechanics.
The most common form of a vector is a collection of components , , where denotes the dimension of the vector space. A function can be consider as a vector in which the discrete index has been replaced by a continuous index . It is useful to introduce a formalism in which this analogy can be exploited without direct reference to the specific forms of the components or .
In Dirac notation, vectors are denoted as . These vectors form a complex linear vector space, which entails the following properties: Any vector can be scaled by any complex number , i.e., we can form new vectors . Furthermore, any two vectors , can be combined into new vectors by a forming a superposition . These operations obey the distributive law . In addition, a vector space possesses a null vector such that , and to each vector there is an inverse vector such that
These properties are all nicely fulfilled for functions. In particular, if and are functions and , are constants, then is also a function.
The scalar product is a generalised versions of the dot product, which associates a complex number to any pair of vectors , . The scalar product fulfils the important property
(116) |
Consistently with this, the scalar product is linear in the second argument, but conjugate linear in the first argument, i.e., , , , .
In general, a scalar product must be positive definite, for . We call the norm of the vector (this generalises the notion of length of an ordinary vector). A vector with is called normalised (this generalises the notion of a unit vector). The procedure of passing from a vector to the normalised vector is called normalisation. Again in analogy to the case of ordinary vectors, two vectors , fulfilling are said to be orthogonal to each other.
In conjunction with a certain completeness condition which is always fulfilled in quantum mechanics, a vector space equipped with a scalar product is called a Hilbert space.
Formally, the scalar product can be interpreted as a product between the vectors and the entities , which form the dual vector space. They represent the left entries in the scalar product and therefore are also conjugate linear: . A dual vector is also called a bra, and an ordinary vector is called a ket, alluding to the fact that in the scalar product they form a bracket (bra-ket). The introduction of these dual vectors is an important step in the Dirac notation; its usefulness will become clear when we discuss operators (generalised matrices).
A basis is a collection of vectors such that any vector can be written as a superposition , where the complex coefficients are unique. The coefficients give a representation of the vector, and can be written as a column vector
(117) |
The corresponding dual vector is written as a row vector . While there are many possible bases, in which the same vector is represented by different coefficients, the number of basis vectors required to obtain all vectors is always the same, and is called the dimension of the vector space ( may be ).
An orthogonal basis fulfills for any . If furthermore for all one speaks of an orthonormal basis. In such a basis, the coefficients representing a vector are given by , and the scalar product takes the explicit form
(118) |
Thus, a vector is normalised if its coefficients in an orthonormal basis obey
(119) |
For functions , the summation over the discrete index is replaced by an integration, . In this case the dimension of the vector space is infinite. The orthonormality of a basis can be stated with help of the Dirac delta function, . In the scalar product, the summation over the discrete index is again replaced by an integration over the continuous index ,
(120) |
This type of integral is called an overlap integral. The expression for the expansion coefficients takes the form , and the normalisation condition translates into
An operator converts any vector into another vector . Linear operators fulfill , where , are complex numbers. Operators can be added according to the rule , and multiplied according to the rule .
In Dirac notation, operators are written as , and the action of an operator is obtained from the multiplication rule . Thus,
(121) |
Assuming that the states , in the definition of form an orthonormal basis, the operator can be represented by -dimensional square matrices
(122) |
where, the matrix elements are obtained from . Since in an orthonormal basis , the operator then acts on a vector according to the standard rules of matrix multiplication, i.e., is represented by a vector with coefficients . Furthermore, the operator addition and multiplication rules then translate to the usual prescriptions of matrix addition and multiplication.
The action of an operator is particularly simple in its eigenrepresentation, defined by a basis fulfilling the eigenvalue equation . The numbers are called eigenvalues, and the associated vectors are called eigenvectors. When appropriate, these eigenvectors are also called eigenfunctions.
If the eigenvectors form an orthonormal basis (as is the case for the hermitian and unitary operators considered below), the eigenrepresentation results in a diagonal matrix, with if and . In Dirac notation, the operator can then be written as .
A particularly simple operator is the identity operator which leaves all states unchanged, . Every state is therefore an eigenstate of , with eigenvalue 1. Consequently, in any orthonormal basis this operator takes the same form . Representations are simply obtained by multiplying out the identities and . In a given orthonormal basis, it is useful to decompose the identity as the sum of projection operators , which fulfill if , and .
For each operator we can define an adjoint operator by setting . In an orthonormal basis we then have . For many operators, we can also define an inverse operator which fulfils .
Two important types of operators are hermitian operators and unitary operators . For any two states , , hermitian operator fulfill , while unitary operators fulfill . This entails and . In an orthonormal basis, the matrix elements of a hermitian operator fulfill , while those of a unitary operator fulfill .
Both classes of operators have the nice property that their sets of normalised eigenvectors form an orthonormal basis. For hermitian operators, the eigenvalues are real, while for unitary operators they fulfill .
Unitary operators are analogous to orthogonal matrices which rotate a coordinate system. In particular, any basis change from one orthonormal basis to another orthonormal basis can be written as , where is a suitable unitary operator. A common form of unitary operators relates them to a hermitian operator via , where is a real constant and the exponential of an operator is defined via its Taylor expansion, . In this case, . Furthermore, the operators and then share the same eigenvectors: If , then with eigenvalues .