Tag Archives: quantum

The Lamb-Dicke limit

The Lamb-Dicke limit is a necessary condition for creation of entangled ions (i.e. the ions must be within the Lamb-Dicke range while their internal and motional states are being manipulated to create the entanglement). The Lamb-Dicke limit defines the upper limit of a range where the ion motion is much smaller than the wavelength of light that is used to excite the desired transition (i.e. the amplitude of the ion motion in the propagation direction of the state manipulating radiation is much less than
λ
/2 Pi, where
λ
is the radiation wavelength). In other words, the Lamb-Dicke limit functionally establishes a maximum temperature for the ions that are to be manipulated. Further, because the ions generally cannot be actively laser cooled while the state manipulations are being performed, the ions must initially be cooled below the Lamb-Dicke limit such that the Lamb-Dicke limit will not be exceeded during the entire manipulation process that creates the entanglement.
π

How do you find the expectation value of an operator in quantum mechanics?

The expected values of position and momentum are given by

\displaystyle \left< x \right>= \int \Psi^* x \Psi\, dx

\displaystyle \left<p\right>=\int \Psi^* \frac{\hbar}{i} \frac{\partial}{\partial x} \Psi\, dx

In general, any dynamic variable, Q, can be expressed in terms of position and momentum, i.e., Q(\hat{p},\hat{x}).

In quantum theory, an experimental setup is described by the observable A to be measured, and the state σ of the system. The expectation value of A in the state σ is denoted as \langle A \rangle_\sigma.

Mathematically, A is a self-adjoint operator on a Hilbert space. In the most commonly used case in quantum mechanics, σ is a pure state, described by a normalized[1] vector ψ in the Hilbert space. The expectation value of A in the state ψ is defined as

(1)       \langle A \rangle_\psi = \langle \psi | A | \psi \rangle .

If dynamics is considered, either the vector ψ or the operator A is taken to be time-dependent, depending on whether the Schrödinger picture or Heisenberg picture is used. The time-dependence of the expectation value does not depend on this choice, however.

If A has a complete set of eigenvectors φj, with eigenvalues aj, then (1) can be expressed as

(2)       \langle A \rangle_\psi = \sum_j a_j |\langle \psi | \phi_j \rangle|^2 .

This expression is similar to the arithmetic mean, and illustrates the physical meaning of the mathematical formalism: The eigenvalues aj are the possible outcomes of the experiment,[2] and their corresponding coefficient |\langle \psi | \phi_j \rangle|^2 is the probability that this outcome will occur; it is often called the transition probability.

A particularly simple case arises when A is a projection, and thus has only the eigenvalues 0 and 1. This physically corresponds to a “yes-no” type of experiment. In this case, the expectation value is the probability that the experiment results in “1”, and it can be computed as

(3)       \langle A \rangle_\psi = \| A \psi \|^2.

References

http://answers.yahoo.com/question/index?qid=20070418194841AAGUVbm

http://en.wikipedia.org/wiki/Expectation_value_(quantum_mechanics)

Introduction to Quantum Mechanics (2nd Edition)

Continue reading

Quantum particle in a box

In quantum mechanics, the particle in a box model (also known as the infinite potential well or the infinite square well) describes a particle free to move in a small space surrounded by impenetrable barriers. The model is mainly used as a hypothetical example to illustrate the differences between classical and quantum systems. In classical systems, for example a ball trapped inside a heavy box, the particle can move at any speed within the box and it is no more likely to be found at one position than another. However, when the well becomes very narrow (on the scale of a few nanometers), quantum effects become important. The particle may only occupy certain positive energy levels. Likewise, it can never have zero energy, meaning that the particle can never “sit still”. Additionally, it is more likely to be found at certain positions than at others, depending on its energy level. The particle may never be detected at certain positions, known as spatial nodes.

The particle in a box model provides one of the very few problems in quantum mechanics which can be solved analytically, without approximations. This means that the observable properties of the particle (such as its energy and position) are related to the mass of the particle and the width of the well by simple mathematical expressions. Due to its simplicity, the model allows insight into quantum effects without the need for complicated mathematics. It is one of the first quantum mechanics problems taught in undergraduate physics courses, and it is commonly used as an approximation for more complicated quantum systems. See also: the history of quantum mechanics.

One-dimensional solution

In quantum mechanics, the wavefunction gives the most fundamental description of the behavior of a particle; the measurable properties of the particle (such as its position, momentum and energy) may all be derived from the wavefunction.[3] The wavefunction ψ(x,t) can be found by solving the Schrödinger equationfor the system

\mathrm{i}\hbar\frac{\partial}{\partial t}\psi(x,t) = -\frac{\hbar^2}{2m}\frac{\partial^2}{\partial x^2}\psi(x,t) +V(x)\psi(x,t),

where \hbar is the reduced Planck constantm is the mass of the particle, i is the imaginary unit and t is time.

Inside the box, no forces act upon the particle, which means that the part of the wavefunction inside the box oscillates through space and time with the same form as a free particle:[1][4]

\psi(x,t) = [A \sin(kx) + B \cos(kx)]\mathrm{e}^{-\mathrm{i}\omega t},\;

where A and B are arbitrary complex numbers. The frequency of the oscillations through space and time are given by the wavenumber k and the angular frequency ω respectively. These are both related to the total energy of the particle by the expression

E = \hbar\omega = \frac{\hbar^2 k^2}{2m},

which is known as the dispersion relation for a free particle.[1]

Initial wavefunctions for the first four states in a one-dimensional particle in a box

The size (or amplitude) of the wavefunction at a given position is related to the probability of finding a particle there by P(x,t) = | ψ(x,t) | 2. The wavefunction must therefore vanish everywhere beyond the edges of the box.[1][4] Also, the amplitude of the wavefunction may not “jump” abruptly from one point to the next.[1] These two conditions are only satisfied by wavefunctions with the form

\psi_n(x,t) = \begin{cases} A \sin(k_n x)\mathrm{e}^{-\mathrm{i}\omega_n t}, & 0 < x < L,\\ 0, & \text{otherwise,} \end{cases}

where n is a positive, whole number. The wavenumber is restricted to certain, specific values given by[5]

k_n = \frac{n \pi}{L}, \quad \mathrm{where} \quad n = \{1,2,3,4,\ldots\},

where L is the size of the box.[7] Negative values of n are neglected, since they give wavefunctions identical to the positive n solutions except for a physically unimportant sign change.[6]

Finally, the unknown constant A may be found by normalizing the wavefunction so that the total probability density of finding the particle in the system is 1. It follows that

\left| A \right| = \sqrt{\frac{2 }{L}}.

Thus, A may be any complex number with absolute value √(2/L); these different values of A yield the same physical state, so A = √(2/L) can be selected to simplify.

Energy levels

The energy of a particle in a box (black circles) and a free particle (grey line) both depend upon wavenumber in the same way. However, the particle in a box may only have certain, discrete energy levels.

The energies which correspond with each of the permitted wavenumbers may be written as[5]

E_n = \frac{n^2\hbar^2 \pi ^2}{2mL^2} = \frac{n^2 h^2}{8mL^2}.

The energy levels increase with n2, meaning that high energy levels are separated from each other by a greater amount than low energy levels are. The lowest possible energy for the particle (its zero-point energy) is found in state 1, which is given by[8]

E_1 = \frac{\hbar^2\pi^2}{2mL^2}.

The particle, therefore, always has a positive energy. This contrasts with classical systems, where the particle can have zero energy by resting motionless at the bottom of the box. This can be explained in terms of the uncertainty principle, which states that the product of the uncertainties in the position and momentum of a particle is limited by

\Delta x\Delta p \geq \frac{\hbar}{2}

It can be shown that the uncertainty in the position of the particle is proportional to the width of the box.[9] Thus, the uncertainty in momentum is roughly inversely proportional to the width of the box.[8] The kinetic energy of a particle is given by Ep2 / (2m), and hence the minimum kinetic energy of the particle in a box is inversely proportional to the mass and the square of the well width, in qualitative agreement with the calculation above.[8]

Spatial location

In classical physics, the particle can be detected anywhere in the box with equal probability. In quantum mechanics, however, the probability density for finding a particle at a given position is derived from the wavefunction as P(x) = | ψ(x) | 2. For the particle in a box, the probability density for finding the particle at a given position depends upon its state, and is given by

P_n(x) = \begin{cases}   \frac{2  }{L}\sin^2\left(\frac{n\pi x}{L}\right); & 0 < x < L \\   0; & \text{otherwise}. \end{cases}

Thus, for any value of n greater than one, there are regions within the box for which P(x) = 0, indicating that spatial nodes exist at which the particle cannot be found.

In quantum mechanics, the average, or expectation value of the position of a particle is given by

\langle x \rangle = \int_{-\infty}^{\infty} \psi^*(x) x \psi(x)\,\mathrm{d}x.

For the particle in a box, it can be shown that the average position is always \langle x \rangle = L/2, regardless of the state of the particle. In other words, the average position at which a particle in a box may be detected is exactly in the center of the quantum well; in agreement with a classical system.

Higher-dimensional boxes

If a particle is trapped in a two-dimensional box, it may freely move in the x and y-directions, between barriers separated by lengths Lx andLy respectively. Using a similar approach to that of the one-dimensional box, it can be shown that the wavefunctions and energies are given respectively by

\psi_{n_x,n_y} = \sqrt{\frac{4}{L_x L_y}} \sin \left( k_{n_x} x \right) \sin \left( k_{n_y} y\right),
E_{n_x,n_y} = \frac{\hbar^2 k_{n_x,n_y}^2}{2m},

where the two-dimensional wavevector is given by

\mathbf{k_{n_x,n_y}} = k_{n_x}\mathbf{\hat{x}} + k_{n_y}\mathbf{\hat{y}} = \frac{n_x \pi }{L_x} \mathbf{\hat{x}} + \frac{n_y \pi }{L_y} \mathbf{\hat{y}}.

For a three dimensional box, the solutions are

\psi_{n_x,n_y,n_z} = \sqrt{\frac{8}{L_x L_y L_z}} \sin \left( k_{n_x} x \right) \sin \left( k_{n_y} y \right) \sin \left( k_{n_z} z \right),
E_{n_x,n_y,n_z} = \frac{\hbar^2 k_{n_x,n_y,n_z}^2}{2m},

where the three-dimensional wavevector is given by

\mathbf{k_{n_x,n_y,n_z}} = k_{n_x}\mathbf{\hat{x}} + k_{n_y}\mathbf{\hat{y}} + k_{n_z}\mathbf{\hat{z}} = \frac{n_x \pi }{L_x} \mathbf{\hat{x}} + \frac{n_y \pi }{L_y} \mathbf{\hat{y}} + \frac{n_z \pi }{L_z} \mathbf{\hat{z}}.

An interesting feature of the above solutions is that when two or more of the lengths are the same (e.g. LxLy), there are multiple wavefunctions corresponding to the same total energy. For example the wavefunction with nx = 2,ny = 1 has the same energy as the wavefunction with nx = 1,ny = 2. This situation is called degeneracy and for the case where exactly two degenerate wavefunctions have the same energy that energy level is said to be doubly degenerate. Degeneracy results from symmetry in the system. For the above case two of the lengths are equal so the system is symmetric with respect to a 90° rotation.

References

http://en.wikipedia.org/wiki/Particle_in_a_box

http://en.wikipedia.org/wiki/Quantum_harmonic_oscillator

Modern Quantum Mechanics (2nd Edition)

Quantum Mechanics Non-Relativistic Theory, Third Edition: Volume 3

Introduction to Quantum Mechanics (2nd Edition)

Introductory Quantum Mechanics (4th Edition)

Stark effect

The Stark effect is the shifting and splitting of spectral lines of atoms and molecules due to the presence of an external static electric field. The amount of splitting and or shifting is called the Stark splitting or Stark shift. In general one distinguishes first- and second-order Stark effects. The first-order effect is linear in the applied electric field, while the second-order effect is quadratic in the field.

The Stark effect is responsible for the pressure broadening (Stark broadening) of spectral lines by charged particles. When the split/shifted lines appear in absorption, the effect is called the inverse Stark effect.

The Stark effect is the electric analogue of the Zeeman effect where a spectral line is split into several components due to the presence of a magnetic field.

The Stark effect can be explained with fully quantum mechanical approaches, but it has also been a fertile testing ground for semiclassical methods.

Mechanism

Classical electrostatics

The Stark effect originates from the interaction between a charge distribution (atom or molecule) and an external electric field. Before turning to quantum mechanics we describe the interaction classically and consider a continuous charge distribution ρ(r). If this charge distribution is non-polarizable its interaction energy with an external electrostatic potential V(r) is

 E_{\mathrm{int}} = \int \rho(\mathbf{r}) V(\mathbf{r}) d\mathbf{r}.\,

If the electric field is of macroscopic origin and the charge distribution is microscopic, it is reasonable to assume that the electric field is uniform over the charge distribution. That is, V is given by a two-term Taylor expansion,

 V(\mathbf{r}) = V(\mathbf{0}) - \sum_{i=1}^3 r_i F_i \quad \hbox{with the electric field:}\quad F_i \equiv  -\left. \left(\frac{\partial V}{\partial r_i} \right)\right|_{\mathbf{0}},

where we took the origin 0 somewhere within ρ. Setting V(\mathbf{0}) as the zero energy, the interaction becomes

  E_{\mathrm{int}} = - \sum_{i=1}^3 F_i  \int \rho(\mathbf{r}) r_i d\mathbf{r} \equiv - \sum_{i=1}^3 F_i  \mu_i = - \mathbf{F}\cdot \boldsymbol{\mu}.

Here we have introduced the dipole moment μ of ρ as an integral over the charge distribution. In case ρ consists of N point charges qj this definition becomes a sum

 \boldsymbol{\mu} \equiv \sum_{j=1}^N  q_j \mathbf{r}_j.

Perturbation theory

Turning now to quantum mechanics we see an atom or a molecule as a collection of point charges (electrons and nuclei), so that the second definition of the dipole applies. The interaction of atom or molecule with a uniform external field is described by the operator

 V_{\mathrm{int}} = - \mathbf{F}\cdot \boldsymbol{\mu}.

This operator is used as a perturbation in first- and second-order perturbation theory to account for the first- and second-order Stark effect.

First order

Let the unperturbed atom or molecule be in a g-fold degenerate state with orthonormal zeroth-order state functions  \psi^0_1, \ldots, \psi^0_g . (Non-degeneracy is the special case g = 1). According to perturbation theory the first-order energies are the eigenvalues of the gg matrix with general element

 (\mathbf{V}_{\mathrm{int}})_{kl} = \langle \psi^0_k |  V_{\mathrm{int}} | \psi^0_l \rangle = -\mathbf{F}\cdot \langle \psi^0_k | \boldsymbol{\mu} | \psi^0_l \rangle, \qquad k,l=1,\ldots, g.

If g = 1 (as is often the case for electronic states of molecules) the first-order energy becomes proportional to the expectation (average) value of the dipole operator \boldsymbol{\mu},

 E^{(1)} = -\mathbf{F}\cdot \langle \psi^0_1 | \boldsymbol{\mu} | \psi^0_1 \rangle = -\mathbf{F}\cdot \langle  \boldsymbol{\mu} \rangle.

Because a dipole moment is a polar vector, the diagonal elements of the perturbation matrix Vint vanish for systems with an inversion center (such as atoms). Molecules with an inversion center in a non-degenerate electronic state do not have a (permanent) dipole and hence do not show a linear Stark effect.

In order to obtain a non-zero matrix Vint for systems with an inversion center it is necessary that some of the unperturbed functions  \psi^0_i have opposite parity (obtain plus and minus under inversion), because only functions of opposite parity give non-vanishing matrix elements. Degenerate zeroth-order states of opposite parity occur for excited hydrogen-like (one-electron) atoms. Such atoms have the principal quantum number n among their quantum numbers. The excited state of hydrogen-like atoms with principal quantum number n is n2-fold degenerate and

 n^2 = \sum_{\ell=0}^{n-1} (2 \ell + 1),

where \ell is the azimuthal (angular momentum) quantum number. For instance, the excited n = 4 state contains the following \ell states,

 16 = 1 + 3 + 5 +7 \;\; \Longrightarrow\;\;  n=4\;\hbox{contains}\; s\oplus p\oplus d\oplus f.

The one-electron states with even \ell are even under parity, while those with odd \ell are odd under parity. Hence hydrogen-like atoms with n>1 show first-order Stark effect.

The first-order Stark effect occurs in rotational transitions of symmetric top molecules (but not for linear and asymmetric molecules). In first approximation a molecule may be seen as a rigid rotor. A symmetric top rigid rotor has the unperturbed eigenstates

 |JKM \rangle = (D^J_{MK})^* \quad\mathrm{with}\quad M,K= -J,-J+1,\dots,J

with 2(2J+1)-fold degenerate energy for |K| > 0 and (2J+1)-fold degenerate energy for K=0. Here DJMK is an element of the Wigner D-matrix. The first-order perturbation matrix on basis of the unperturbed rigid rotor function is non-zero and can be diagonalized. This gives shifts and splittings in the rotational spectrum. Quantitative analysis of these Stark shift yields the permanent electric dipole moment of the symmetric top molecule.

Second order

As stated, the quadratic Stark effect is described by second-order perturbation theory. The zeroth-order problems

 H^{(0)} \psi^0_k = E^{(0)}_k \psi^0_k, \quad k=0,1, \ldots, \quad E^{(0)}_0 < E^{(0)}_1 \le E^{(0)}_2, \dots

are assumed to be solved. It is usual to assume that the zeroth-order state to be perturbed is non-degenerate. If we take the ground state as the non-degenerate state under consideration (for hydrogen-like atoms: n = 1), perturbation theory gives

 E^{(2)} = \sum_{k>0} \frac{\langle \psi^0_0 | V_\mathrm{int} | \psi^0_k \rangle \langle \psi^0_k | V_\mathrm{int} | \psi^0_0 \rangle}{E^{(0)}_0 - E^{(0)}_k} =- \frac{1}{2} \sum_{i,j=1}^3 F_i \alpha_{ij} F_j

with the components of the polarizability tensor α defined by

 \alpha_{ij}\equiv -2\sum_{k>0} \frac{\langle \psi^0_0 | \mu_i | \psi^0_k \rangle \langle \psi^0_k | \mu_j | \psi^0_0\rangle}{E^{(0)}_0 - E^{(0)}_k}.

The energy E(2) gives the quadratic Stark effect.

Because of their spherical symmetry the polarizability tensor of atoms is isotropic,

 \alpha_{ij} = \alpha_0 \delta_{ij} \Longrightarrow E^{(2)} = -\frac{1}{2} \alpha_0 F^2,

which is the quadratic Stark shift for atoms. For many molecules this expression is not too bad an approximation, because molecular tensors are often reasonably isotropic.

Problems

The perturbative treatment of the Stark effect has some problems. In the presence of an electric field, states of atoms and molecules that were previously bound (square-integrable), become formally (non-square-integrable) resonances of finite width. These resonances may decay in finite time via field ionization. For low lying states and not too strong fields the decay times are so long, however, that for all practical purposes the system can be regarded as bound. For highly excited states and/or very strong fields ionization may have to be accounted for. (See also the article on the Rydberg atom).

References

Modern Quantum Mechanics (2nd Edition)

http://en.wikipedia.org/wiki/Stark_effect

The Theory of Atomic Spectra

Angular momentum algebra

Canonical commutation relation

the canonical commutation relation is the relation between canonical conjugate quantities (quantities which are related by definition such that one is the Fourier transform of another), for example:

[x,p_x] = i\hbar

between the position x and momentum px in the x direction of a point particle in one dimension, where [x,px] = xpx − pxx is the commutator of x and pxi is the imaginary unit, and ħ is the reducedPlanck’s constant h /2π . This relation is attributed to Max Born, and it was noted by E. Kennard (1927) to imply the Heisenberg uncertainty principle.

By contrast, in classical physics, all observables commute and the commutator would be zero. However, an analogous relation exists, which is obtained by replacing the commutator with the Poisson bracket multiplied by i ħ:

\{x,p\} = 1 \,      .

This observation led Dirac to propose that the quantum counterparts \hat f,\hat g of classical observables fg satisfy

[\hat f,\hat g]= i\hbar\widehat{\{f,g\}} \, .

According to the standard mathematical formulation of quantum mechanics, quantum observables such as x and p should be represented as self-adjoint operators on some Hilbert space. It is relatively easy to see that two operators satisfying the canonical commutation relations cannot both be bounded. The canonical commutation relations can be made tamer by writing them in terms of the (bounded) unitary operators e − ikx and e − iap. The result is the so-called Weyl relations. The uniqueness of the canonical commutation relations between position and momentum is guaranteed by theStone-von Neumann theorem. The group associated with the commutation relations is called the Heisenberg group.

Angular momentum operators

 [{L_x}, {L_y}] = i \hbar \epsilon_{xyz} {L_z},

where εxyz is the Levi-Civita symbol and simply reverses the sign of the answer under pairwise interchange of the indices. An analogous relation holds for the spin operators.

All such nontrivial commutation relations for pairs of operators lead to corresponding uncertainty relations (H. P. Robertson[2]), involving positive semi-definite expectation contributions by their respective commutators and anticommutators. In general, for two Hermitian operators A and B, consider expectation values in a system in the state ψ, the variances around the corresponding expectation values being (ΔA)2 ≡ 〈 (A −<A>)2 〉, etc.

Then

 \Delta  A \, \Delta  B \geq  \frac{1}{2} \sqrt{ \left|\left\langle\left[{A},{B}\right]\right\rangle \right|^2 + \left|\left\langle\left\{ A-\langle A\rangle ,B-\langle B\rangle  \right\} \right\rangle \right|^2} ,

where [A,B] ≡ ABBA is the commutator of A and B, and {A,B} ≡ AB+BA is the anticommutator. This follows through use of the Cauchy–Schwarz inequality, since |〈A2〉| |〈B2〉| ≥ |〈AB〉|2, and AB = ([A,B] + {A,B}) /2 ; and similarly for the shifted operators A−〈A〉 and B−〈B〉 . Judicious choices for A and B yield Heisenberg’s familiar uncertainty relation, for x and p, as usual; or, here, Lx and Ly , in angular momentum multiplets, ψ = |lm 〉 , useful constraints such as l (l+1) ≥ m (m+1), and hence l ≥ m, among others.

References

Lie Groups for Pedestrians

Modern Quantum Mechanics (2nd Edition)

http://www.dfcd.net/articles/firstyear/lectures/angmom.pdf

http://en.wikipedia.org/wiki/Canonical_commutation_relation

http://galileo.phys.virginia.edu/classes/751.mf1i.fall02/AngularMomentum.htm