1
Chapter 8. Chemical Dynamics
Chemical dynamics is a field in which scientists study the rates and mechanisms
of chemical reactions. It also involves the study of how energy is transferred among
molecules as they undergo collisions in gas-phase or condensed-phase environments.
Therefore, the experimental and theoretical tools used to probe chemical dynamics must
be capable of monitoring the chemical identity and energy content (i.e., electronic,
vibrational, and rotational state populations) of the reacting species. Moreover, because
the rates of chemical reactions and energy transfer are of utmost importance, these tools
must be capable of doing so on time scales over which these processes, which are often
very fast, take place. Let us begin by examining many of the most commonly employed
theoretical models for simulating and understanding the processes of chemical dynamics.
I. Theoretical Tools for Studying Chemical Change and Dynamics
A. Transition State Theory
The most successful and widely employed theoretical approach for studying reaction
rates involving species that are undergoing reaction at or near thermal-equilibrium
conditions is the transition state theory (TST) of Eyring. This would not be a good way to
model, for example, photochemical reactions in which the reactants do not reach thermal
2
equilibrium before undergoing significant reaction progress. However, for most thermal
reactions, it is remarkably successful.
In this theory, one views the reactants as undergoing collisions that act to keep all of
their degrees of freedom (translational, rotational, vibrational, electronic) in thermal
equilibrium. Among the collection of such reactant molecules, at any instant of time,
some will have enough internal energy to access a transition state (TS) on the Born-
Oppenheimer ground state potential energy surface. Within TST, the rate of progress
from reactants to products is then expressed in terms of the concentration of species that
exist near the TS multiplied by the rate at which these species move through the TS
region of the energy surface.
The concentration of species at the TS is, in turn, written in terms of the equilibrium
constant expression of statistical mechanics discussed in Chapter 7. For example, for a
bimolecular reaction A + B → C passing through a TS denoted AB*, one writes the
concentration (in molecules per unit volume) of AB* species in terms of the
concentrations of A and of B and the respective partition functions as
[AB*] = (qAB*/V)/{(qA/V)( qB/V)} [A] [B].
3
There is, however, one aspect of the partition function of the TS species that is specific to
this theory. The qAB* contains all of the usual translational, rotational, vibrational, and
electronic partition functions that one would write down, as we did in Chapter 7, for a
conventional AB molecule except for one modification. It does not contain a {exp(-hνj
/2kT)/(1- exp(-hνj/kT))} vibrational contribution for motion along the one internal
coordinate corresponding to the reaction path.
Figure 8.1 Typical Potential Energy Surface in Two Dimensions Showing Local Minima,
Transition States and Paths Connecting Them.
In the vicinity of the TS, the reaction path can be identified as that direction along which
the PES has negative curvature; along all other directions, the energy surface is positively
curved. For example, in Fig. 8.1, a reaction path begins at Transition Structure B and is
directed "downhill". More specifically, if one knows the gradients {(?E/?sk) }and
4
Hessian matrix elements { Hj,k = ?2E/?sj?sk}of the energy surface at the TS, one can
express the variation of the potential energy along the 3N Cartesian coordinates {sk} of
the molecule as follows:
E (sk) = E(0) + Σk (?E/?sk) sk + 1/2 Σj,k sj Hj,k sk + …
where E(0) is the energy at the TS, and the {sk} denote displacements away from the TS
geometry. Of course, at the TS, the gradients all vanish because this geometry
corresponds to a stationary point. As we discussed in the Background Material, the
Hessian matrix Hj,k has 6 zero eigenvalues whose eigenvectors correspond to overall
translation and rotation of the molecule. This matrix has 3N-7 positive eigenvalues whose
eigenvectors correspond to the vibrations of the TS species, as well as one negative
eigenvalue. The latter has an eigenvector whose components {sk} along the 3N Cartesian
coordinates describe the direction of the reaction path as it begins its journey from the TS
backward to reactants (when followed in one direction) and onward to products (when
followed in the opposite direction). Once one moves a small amount along the direction
of negative curvature, the reaction path is subsequently followed by taking infinitesimal
“steps” downhill along the gradient vector g whose 3N components are (?E/?sk). Note
that once one has moved downhill away from the TS by taking the initial step along the
negatively curved direction, the gradient no longer vanishes because one is no longer at
the stationary point.
Returning to the TST rate calculation, one therefore is able to express the
concentration [AB*] of species at the TS in terms of the reactant concentrations and a
5
ratio of partition functions. The denominator of this ratio contains the conventional
partition functions of the reactant molecules and can be evaluated as discussed in Chapter
7. However, the numerator contains the partition function of the TS species but with one
vibrational component missing (i.e., qvib = Πk=1,3N-7 {exp(-hνj /2kT)/(1- exp(-hνj/kT))}).
Other than the one missing qvib, the TS's partition function is also evaluated as in Chapter
7. The motion along the reaction path coordinate contributes to the rate expression in
terms of the frequency (i.e., how often) with which reacting flux crosses the TS region
given that the system is in near-thermal equilibrium at temperature T.
To compute the frequency with which trajectories cross the TS and proceed
onward to form products, one imagines the TS as consisting of a narrow region along the
reaction coordinate s; the width of this region we denote δs. We next ask what the
classical weighting factor is for a collision to have momentum ps along the reaction
coordinate. Remembering our discussion of such matters in Chapter 7, we know that the
momentum factor entering into the classical partition function for translation along the
reaction coordinate is (1/h) exp(-ps2/2μkT) dps. Here, μ is the mass factor associated with
the reaction coordinate s. We can express the rate or frequency at which such trajectories
pass through the narrow region of width δs as (ps/μδs), with ps/μ being the speed of
passage (cm s-1) and 1/δs being the inverse of the distance that defines the TS region. So,
(ps/μδs) has units of s-1. In summary, we expect the rate of trajectories moving through
the TS region to be
(1/h) exp(-ps2/2μkT) dps (ps/μδs).
6
However, we still need to integrate this over all values of ps that correspond to enough
energy ps2/2μ to access the TS’s energy, which we denote E*. Moreover, we have to
account for the fact that it may be that not all trajectories with kinetic energy equal to E*
or greater pass on to form product molecules; some trajectories may pass through the TS
but later recross the TS and return to produce reactants. Moreover, it may be that some
trajectories with kinetic energy along the reaction coordinate less than E* can react by
tunneling through the barrier.
The way we account for the facts that a reactive trajectory must have at least E* in
energy along s and that not all trajectories with this energy will react is to integrate over
only values of ps greater than (2μE*)1/2 and to include in the integral a so-called
transmission coefficient κ that specifies the fraction of trajectories crossing the TS that
eventually proceed onward to products. Putting all of these pieces together, we carry out
the integration over ps just described to obtain:
∫ ∫ (1/h) κ exp(-ps2/2μkT) (ps/μδs) ds dps
where the momentum is integrated from ps = (2μE*)1/2 to ∞ and the s-coordinate is
integrated only over the small region δs. If the transmission coefficient is factored out of
the integral (treating it as a multiplicative factor), the integral over ps can be done and
yields the following:
κ (kT/h) exp(-E*/kT).
7
The exponential energy dependence is usually then combined with the partition function
of the TS species that reflect this species’ other 3N-7 vibrational coordinates and
momenta and the reaction rate is then expressed as
Rate = κ (kT/h) [AB*] = κ (kT/h) (qAB* /V)/{(qA/V)( qB/V)} [A] [B].
This implies that the rate coefficient krate for this bimolecular reaction is given in terms of
molecular partition functions by:
krate = κ kT/h (qAB*/V)/{(qA/V)(qB/V)}
which is the fundamental result of TST. Once again we notice that ratios of partition
functions per unit volume can be used to express ratios of species concentrations (in
number of molecules per unit volume), just as appeared in earlier expressions for
equilibrium constants as in Chapter 7.
The above rate expression undergoes only minor modifications when
unimolecular reactions are considered. For example, in the hypothetical reaction A → B
via the TS (A*), one obtains
krate = κ kT/h {(qA*/V)/(qA/V)},
where again qA* is a partition function of A* with one missing vibrational component.
8
Before bringing this discussion of TST to a close, I need to stress that this theory is
not exact. It assumes that the reacting molecules are nearly in thermal equilibrium, so it is
less likely to work for reactions in which the reactant species are prepared in highly non-
equilibrium conditions. Moreover, it ignores tunneling by requiring all reactions to
proceed through the TS geometry. For reactions in which a light atoms (i.e., an H or D
atom) is transferred, tunneling can be significant, so this conventional form of TST can
provide substantial errors in such cases. Nevertheless, TST remains the most widely used
and successful theory of chemical reaction rates and can be extended to include tunneling
and other corrections as we now illustrate.
B. Variational Transition State Theory
Within the TST expression for the rate constant of a bi-molecular reaction, krate = κ
kT/h (qAB*/V)/{(qA/V)(qB/V)}or of a uni-molecular reaction, krate = κ kT/h
{(qA*/V)/(qA/V)}, the height (E*) of the barrier on the potential energy surface
appears in the TS species’ partition function qAB* or qA*, respectively. In particular,
the TS partition function contains a factor of the form exp(-E*/kT) in which the Born-
Oppenheimer electronic energy of the TS relative to that of the reactant species
appears. This energy E* is the value of the potential energy E(S) at the TS geometry,
which we denote S0.
It turns out that the conventional TS approximation to krate over-estimates reaction
rates because it assumes all trajectories that cross the TS proceed onward to products
unless the transmission coefficient is included to correct for this. In the variational
transition state theory (VTST), one does not evaluate the ratio of partition functions
9
appearing in krate at S0, but one first determines at what geometry (S*) the TS partition
function (i.e., qAB* or qA*) is smallest. Because this partition function is a product of
(i) the exp(-E(S)/kT) factor as well as (ii) 3 translational, 3 rotational, and 3N-7
vibrational partition functions (which depend on S), the value of S for which this
product is smallest need not be the conventional TS value S0. What this means is that
the location (S*) along the reaction path at which the free-energy reaches a saddle
point is not the same the location S0 where the Born-Oppenheimer electronic energy
E(S) has its saddle. This interpretation of how S* and S0 differ can be appreciated by
recalling that partition functions are related to the Helmholtz free energy A by q =
exp(-A/kT); so determining the value of S where q reaches a minimum is equivalent
to finding that S where A is at a maximum.
So, in VTST, one adjusts the “dividing surface” (through the location of the
reaction coordinate S) to first find that value S* where krate has a minimum. One then
evaluates both E(S*) and the other components of the TS species partition functions
at this value S*. Finally, one then uses the krate expressions given above, but with S
taken at S*. This is how VTST computes reaction rates in a somewhat different
manner than does the conventional TST. As with TST, the VTST, in the form
outlined above, does not treat tunneling and the fact that not all trajectories crossing
S* proceed to products. These corrections still must be incorporated as an “add-on” to
this theory (i.e., in the κ factor) to achieve high accuracy for reactions involving light
species (recall from the Background Material that tunneling probabilities depend
exponentially on the mass of the tunneling particle).
10
C. Reaction Path Hamiltonian Theory
Let us review what the reaction path is as defined above. It is a path that
i. begins at a transition state (TS) and evolves along the direction of negative curvature on
the potential energy surface (as found by identifying the eigenvector of the Hessian
matrix Hj,k = ?2E/?sk?sj that belongs to the negative eigenvalue);
ii. moves further downhill along the gradient vector g whose components are gk = ?E/?sk’
iii. terminates at the geometry of either the reactants or products (depending on whether
one began moving away from the TS forward or backward along the direction of negative
curvature).
The individual “steps” along the reaction coordinate can be labeled S0, S1, S2, … SP as
they evolve from the TS to the products (labeled SP) and S-R, S-R+1, …S0 as they evolve
from reactants (S-R) to the TS. If these steps are taken in very small (infinitesimal)
lengths, they form a continuous path and a continuous coordinate that we label S.
At any point S along a reaction path, the Born-Oppenheimer potential energy
surface E(S), its gradient components gk(S) = (?E(S)/?sk) and its Hessian components
Hk,j(S) = (?2E(S)/?sk?sj) can be evaluated in terms of derivatives of E with respect to the
3N Cartesian coordinates of the molecule. However, when one carries out reaction path
dynamics, one uses a different set of coordinates for reasons that are similar to those that
arise in the treatment of normal modes of vibration as given in the Background Material.
In particular, one introduces 3N mass-weighted coordinates xj = sj (mj)1/2 that are related
to the 3N Cartesian coordinates sj in the same way as we saw in the Background Material.
11
The gradient and Hessian matrices along these new coordinates {xj} can be
evaluated in terms of the original Cartesian counterparts:
gk’(S) = gk(S) (mk)-1/2
Hj,k’ = Hj,k (mjmk)-1/2.
The eigenvalues {ωk2} and eigenvectors {vk} of the mass-weighted Hessian H’ can then
be determined. Upon doing so, one finds
i. 6 zero eigenvalues whose eigenvectors describe overall rotation and translation of
the molecule;
ii. 3N-7 positive eigenvalues {ωK2} and eigenvectors vK along which the gradient g
has zero (or nearly so) components;
iii. and one eigenvalue ωS2 (that may be positive, zero, or negative) along whose
eigenvector vS the gradient g has its largest component.
The one unique direction along vS gives the direction of evolution of the reaction path (in
these mass-weighted coordinates). All other directions (i.e., within the space spanned by
the 3N-7 other vectors {vK}) possess zero gradient component and positive curvature.
This means that at any point S on the reaction path being discussed
i. one is at a local minimum along all 3N-7 directions {vK} that are transverse to the
reaction path direction (i.e., the gradient direction);
ii. one can move to a neighboring point on the reaction path by moving a small
(infinitesimal) amount along the gradient.
12
In terms of the 3N-6 mass-weighted Hessian’s eigen-mode directions ({vK} and
vS), the potential energy surface can be approximated, in the neighborhood of each such
point on the reaction path S, by expanding it in powers of displacements away from this
point. If these displacements are expressed as components
i. δXk along the 3N-7 eigenvectors vK and
ii. δS along the gradient direction vS,
one can write the Born-Oppenheimer potential energy surface locally as:
E = E(S) + vS δS + 1/2 ωS2 δS2 + ΣK=1,3N-7 1/2 ωK2 δXK2 .
Within this local quadratic approximation, E describes a sum of harmonic potentials
along each of the 3N-7 modes transverse to the reaction path direction. Along the
reaction path, E appears with a non-zero gradient and a curvature that may be positive,
negative, or zero.
The eigenmodes of the local (i.e., in the neighborhood of any point S along the
reaction path) mass-weighted Hessian decompose the 3N-6 internal coordinates into 3N-7
along which E is harmonic and one (S) along which the reaction evolves. In terms of
these same coordinates, the kinetic energy T can also be written and thus the classical
Hamiltonian H = T + V can be constructed. Because the coordinates we use are mass-
weighted, in Cartesian form the kinetic energy T contains no explicit mass factors:
T = 1/2 Σj mj (dsj/dt)2 = 1/2 Σj (dxj/dt)2.
13
This means that the momenta conjugate to each (mass-weighted) coordinate xj, obtained
in the usual way as pj = ?[T-V]/?(dxj/dt) = dxj/dt, all have identical (unit) mass factors
associated with them.
To obtain the working expression for the reaction path Hamiltonian (RPH), one
must transform the above equation for the kinetic energy T by replacing the 3N Cartesian
mass-weighted coordinates {xj} by
i. the 3N-7 eigenmode displacement coordinates δXj,
ii. the reaction path displacement coordinate δS, and
iii. 3 translation and 3 rotational coordinates.
The 3 translational coordinates can be separated and ignored (because center-of-mass
energy is conserved) in further consideration. The 3 rotational coordinates do not enter
into the potential E, but they do appear in T. However, it is most common to ignore their
effects on the dynamics that occurs in the internal-coordinates; this amounts to ignoring
the effects of overall centrifugal forces on the reaction dynamics. We will proceed with
this approximation in mind.
Although it is tedious to perform the coordinate transformation of T outlined
above, it has been done and results in the following form for the RPH:
H = ΣK=1,3N-7 1/2[pK2 + ωK2(S)] + E(S) + 1/2 [pS - ΣK,K’=1,3N-7 pK δXK’ BK,K’]2 /(1+F)
where
14
(1+F) = [1 + ΣK=1,3N-7 δXK BK,S]2.
In the absence of the so-called dynamical coupling factors BK,K’ and BK,S, this expression
for H describes
(1) 3N-7 harmonic-oscillator Hamiltonia 1/2[pK2 + ωK2(S)] each of which has a locally
defined frequency ωK(S) that varies along the reaction path (i.e., is S-dependent);
(2) a Hamiltonian 1/2 pS2 + E(S) for motion along the reaction coordinate S with E(S)
serving as the potential.
In this limit (i.e., with the B factors “turned off”), the reaction dynamics can be simulated
in what is termed a “vibrationally adiabatic” manner by
i. placing each transverse oscillator into a quantum level vK that characterizes the
reactant’s population of this mode;
ii. assigning an initial momentum pS(0) to the reaction coordinate that is
characteristic of the collision to be simulated (e.g., pS(0) could be sampled from a
Maxwell-Boltzmann distribution if a thermal reaction is of interest, or pS(0) could
be chosen equal to the mean collision energy of a beam-collision experiment);
iii. time evolving the S and pS, coordinate and momentum using the above
Hamiltonian, assuming that each transverse mode remains in the quantum state vK
that it had when the reaction began.
The assumption that vK remains fixed, which is why this model is called vibrationally
adiabatic, does not mean that the energy content of the Kth mode remains fixed because
the frequencies ωK(S) vary as one moves along the reaction path. As a result, the kinetic
15
energy along the reaction coordinate 1/2 pS2 will change both because E(S) varies along S
and because ΣK=1,3N-7 hωK2(S) [vK + 1/2] varies along S.
Let’s return now to the RPH theory in which the dynamical couplings among
motion along the reaction path and the modes transverse to it are included. In the full
RPH, the terms BK,K’(S) couple modes K and K’, while BK,S(S) couples the reaction path
to mode K. These couplings are how energy can flow among these various degrees of
freedom. Explicit forms for the BK,K’ and BK,S factors are given in terms of the
eigenvectors {vK, vS} of the mass-weighted Hessian matrix as follows:
BK,K’ = <dvK/dS| vK’>; BK,S = <dvK/dS | vS>
where the derivatives of the eigenvectors {dvK/dS} are usually computed by taking the
eigenvectors at two neighboring points S and S’ along the reaction path:
dvK/dS = {vK(S’) – vK(S))/(S’-S).
In summary, once a reaction path has been mapped out, one can compute, along
this path, the mass-weighted Hessian matrix and the potential E(S). Given these
quantities, all terms in the RPH
H = ΣK=1,3N-7 1/2[pK2 + ωK2(S)] + E(S) + 1/2 [pS - ΣK,K’=1,3N-7 pK δXK’ BK,K’]2 /(1+F)
are in hand.
16
This knowledge can, subsequently, be used to perform the propagation of a set of
classical coordinates and momenta forward in time. For any initial (i.e., t = 0) momenta
pS and pK , one can use the above form for H to propagate the coordinates {δXK, δS} and
momenta {pK, pS} forward in time. In this manner, one can use the RPH theory to follow
the time evolution of a chemical reaction that begins (t = 0) with coordinates and moment
characteristic of reactants under specified laboratory conditions and moves through a TS
and onward to products. Once time has evolved long enough for product geometries to be
realized, one can interrogate the values of 1/2[pK2 + ωK2(S)] to determine how much
energy has been deposited into various product-molecule vibrations and of 1/2 pS2 to see
what the final kinetic energy of the product fragments is. Of course, one also monitors
what fraction of the trajectories, whose initial conditions are chosen to represent some
experimental situation, progress to product geometries vs. returning to reactant
geometries. In this way, one can determine the overall reaction probability.
D. Classical Dynamics Simulation of Rates
One can perform classical dynamics simulations of reactive events without using
the reaction path Hamiltonian. Following a procedure like that outlined in Chapter 7
where condensed-media MD simulations were discussed, one can time-evolve the
Newton equations of motion of the molecular reaction species using, for example, the
Cartesian coordinates of each atom in the system and with either a Born-Oppenheimer
surface or a parameterized functional form. Of course, it is essential for whatever
17
function one uses to accurately describe the reactive surface, especially near the transition
state.
With each such coordinate having an initial velocity (dq/dt)0 and an initial value
q0 , one then uses Newton’s equations written for a time step of duration δt to propagate q
and dq/dt forward in time according, for example , to the following first-order
propagation formula:
q(t+δ t) = q0 + (dq/dt)0 δt
dq/dt (t+δt) = (dq/dt)0 - δt [(?E/?q)0/mq].
Here mq is the mass factor connecting the velocity dq/dt and the momentum pq conjugate
to the coordinate q:
pq = mq dq/dt,
and -(?E/?q)0 is the force along the coordianate q at the “initial” geometry q0. Again, as
in Chapter 7, I should note that the above formulas for propagating q and p forward in
time represent only the most elementary approach to this problem. There are other, more
sophisticated, numerical methods for effecting more accurate and longer-time
propagations, but I will not go into them here. Rather, I wanted to focus on the basics of
how these simulations are carried out.
18
By applying the time-propagation proecess, one generates a set of “new”
coordinates q(t+δt) and new velocities dq/dt(t+δt) appropriate to the system at time t+δt.
Using these new coordinates and momenta as q0 and (dq/dt)0 and evaluating the forces
–(?E/?q)0 at these new coordinates, one can again use the Newton equations to generate
another finite-time-step set of new coordinates and velocities. Through the sequential
application of this process, one generates a sequence of coordinates and velocities that
simulate the system’s dynamical behavior.
In using this kind of classical trajectory approach to study chemical reactions, it is
important to choose the initial coordinates and momenta in a way that is representative of
the experimental conditions that one is attempting to simulate. The tools of statistical
mechanics discussed in Chapter 7 guide us in making these choices. When one attempts,
for example, to simulate the reactive collisions of an A atom with a BC molecule to
produce AB + C, it is not appropriate to consider a single classical (or quantal) collision
between A and BC. Why? Because in any laboratory setting,
1. The A atoms are probably moving toward the BC molecules with a distribution of
relative speeds. That is, within the sample of molecules (which likely contains 1010 or
more molecules), some A + BC pairs have low relative kinetic energies when they
collide, and others have higher relative kinetic energies. There is a probability
distribution P(EKE ) for this relative kinetic energy that must be properly sampled in
choosing the initial conditions.
2. The BC molecules may not all be in the same rotational (J) or vibrational (v) state.
There is a probability distribution function P(J,v) describing the fraction of BC molecules
that are in a particular J state and a particular v state. Initial values of the BC molecule's
19
internal vibrational coordinate and momentum as well as its orientation and rotational
angular momentum must be selected to represent this P(J,v).
3. When the A and BC molecules collide with a relative motion velocity vector v, they do
not all hit "head on". Some collisions have small impact parameter b (the closest distance
from A to the center of mass of BC if the collision were to occur with no attractive or
repulsive forces), and some have large b-values (see Fig. 8.2). The probability function
for these impact parameters is P(b) = 2pi b db, which is simply a statement of the
geometrical fact that larger b-values have more geometrical volume element than smaller
b-values.
Figure 8.2 Coordinates Needed to Characterize an Atom-Diatom Collision Showing the
Impact Parameter b.
20
So, to simulate the entire ensemble of collisions that occur between A atoms and
BC molecules in various J, v states and having various relative kinetic energies EKE and
impact parameters b, one must:
1. run classical trajectories (or quantum propagations) for a large number of J, v, EKE ,
and b values,
2. with each such trajectory assigned an overall weighting (or importance) of
Ptotal = P(EKE ) P(J,v) 2pib db.
After such an ensemble of trajectories representative of an experimental condition
has been carried out, one has available a great deal of data. This data includes knowledge
of what fraction of the trajectories produced final geometries characteristic of products,
so the net reaction probability can be calculated. In addition, the kinetic and potential
energy content of the internal (vibrational and rotational) modes of the product molecules
can be interrogated and used to compute probabilities for observing products in these
states. This is how classical dynamics simulations allow us to study chemical reactions
and/or energy transfer.
E. RRKM Theory
Another theory that is particularly suited for studying uni-molecular decomposition
reactions is named after the four scientists who developed it- Rice, Ramsperger, Kassel,
and Marcus. To use this theory, one imagines an ensemble of molecules that have been
21
“activated” to a state in which they possess a specified total amount of internal energy E
of which an amount E*rot exists as rotational energy and the remainder as internal
vibrational energy.
The mechanism by which the molecules become activated could involve collisions
or photochemistry. It does not matter as long as enough time has passed to permit one to
reasonably assume that these molecules have the energy E-E*rot distributed randomly
among all their internal vibrational degrees of freedom. When considering thermally
activated unimolecular decomposition of a molecule, the implications of such
assumptions are reasonably clear. For photochemically activated unimolecular
decomposition processes, one usually also assumes that the molecule has undergone
radiationless relaxation and returned to its ground electronic state but in a quite
vibrationally “hot” situation. That is, in this case, the molecule contains excess
vibrational energy equal to the energy of the optical photon used to excite it. Finally,
when applied to bimolecular reactions, one assumes that collision between the two
fragments results in a long-lived complex. The lifetime of this intermediate must be long
enough to allow the energy E-E*rot, which is related to the fragments’ collision energy, to
be randomly distributed among all vibrational modes of the collision complex.
For bimolecular reactions that proceed “directly” (i.e., without forming a long-lived
intermediate), one does not employ RRKM-type theories because their primary
assumption of energy randomization almost certainly would not be valid in such cases.
The RRKM expression of the unimolecular rate constant for activated molecules A*
(i.e., either a long-lived complex formed in a bimolecular collision or a “hot” molecule)
dissociating to products through a transition state, A* → TS → P, is
22
krate = G(E-E0 –E’tot )/(N(E-E*rot) h).
Here, the total energy E is related to the energies of the activated molecules by
E = E*rot + E*vib
where E*rot is the rotational energy of the activated molecule and E*vib is the vibrational
energy of this molecule. This same energy E must, of course, appear in the transition state
where it is decomposed as an amount E0 needed to move from A* to the TS (i.e, the
energy needed to reach the barrier) and vibrational (E'vib) , translational (E'trans along the
reaction coordiate), and rotational (E'rot) energies:
E = E0 + E’vib + E’trans + E’rot .
In the rate coefficient expression, G(E-E0 –E’rot ) is the total sum of internal
vibrational quantum states that the transition state possesses having energies up to and
including E-E0 –E’rot . This energy is the total energy E but with the activation energy E0
removed and the overall rotational energy E’rot of the TS removed. The quantity
N(E-E*rot) is the density of internal vibrational quantum states (excluding the mode
describing the reaction coordinate) that the activated molecule possesses having an
energy between E-E*rot and E-E*rot + dE. In this expression, the energy E-E*rot is the total
energy E with the rotational energy E*rot of the activated species removed.
23
In the most commonly employed version of RRKM theory, the rotational energies of
the activated molecules E*rot and of the TS E’rot are assumed to be related by
E*rot – E’rot = J(J+1) h2/8pi2 {1/I* - 1/I’} = E*rot {1 – I*/I’}.
Here I* and I’ are the average (taken over the three eigenvalues of the moment inertia
tensors) moments of inertia of the activated molecules and TS species, respectively. The
primary assumption embodied in the above relationship is that the rotational angular
momenta of the activated and TS species are the same, so their rotational energies can be
related, as expressed in the equation, to changes in geometries as reflected in their
mometns of inertia. Because RRKM theory assumes that the vibrational energy is
randomly distributed, its fundamental rate coefficient equation
krate = G(E-E0 –E’tot )/(N(E-E*rot) h) depends on the total energy E, the energy E0 required
to access the TS, and the amount of energy contained in the rotational degrees of freedom
that is thus not available to the vibrations.
To implement a RRKM rate coefficient calculation, one must know
(i) the total energy E available,
(ii) the barrier energy E0,
(iii) the geometries (and hence the moments of inertia I* and I’) of the activated
molecules and of the TS, respectively,
(iv) the rotational energy E*rot of the activated molecules, as well as
(v) all 3N-6 vibrational energies of the activated molecules and all 3N-7 vibrational
energiers of the TS (i.e., excluding the reaction coordinate).
24
The rotational energy of the TS species can then be related to that of the activated
molecules through E*rot – E’rot = E*rot {1 – I*/I’}.
To simulate an experiment in which the activated molecules have a thermal
distribution of rotational energies, the RRKM rate constant is computed for a range of
E*rot values and then averaged over E*rot using the thermal Boltzmann population
(2J+1) exp(-J(J+1)h2/(8pi2I*kT)) as a weighting factor. This can be carried out, for
example, using the MC process for selecting rotational J values. This then produces a rate
constant for any specified total energy E. Alternatively, to simulate experiments in which
the activated species are formed in bimolecular collisions at a specified energy E, the
RRKM rate coefficient is computed for a range of E*rot values with each E*rot related to
the collisional impact parameter b that we discussed earlier. In that case, the collisional
angular momentum J is given as J = μ v b, where v is the collision speed (related to the
collision energy) and μ is the reduced mass of the two colliding fragments. Again using
E*rot – E’rot = E*rot {1 – I*/I’} the TS rotational energy can be related to that of the
activated species. Finally, the RRKM rate coefficient is evaluated by averaging the result
over a series of impact parameters b (each of which implies a J value and thus an E*rot )
with 2pib db as the weighting factor.
The evaluation of the sum of states G(E-E0 –E’tot ) and the density of states
N(E-E*rot) that appear in the RRKM expression is usually carried out using a state-
counting algorithm such as that implemented by Beyer and Swinehart (Commun. Assoc.
Comput. Machin. 16, 372 (1973)). This algorithm uses knowledge of the 3N-6 harmonic
vibrational frequencies of the activated molecules and the 3N-7 frequencies of the TS and
determines how many ways a given amount of energy can be distributed among these
25
modes. By summing over all such distributions for energy varying from zero to E, the
algorithm determines G(E). By taking the difference G(E+δE) – G(E), it determines
N(E)δE.
F. Correlation Function Expressions for Rates
Recall from Chapter 6 that rates of photon absorption can, in certain
circumstances, be expressed either (a) in terms of squares of transition dipole matrix
elements connecting each initial state Φi to each final state Φf,
| E0 ? <Φf | μ | Φi> |2
or (b) in terms of the equilibrium average of the product of a transition dipole vector at
time t=0 dotted into this same vector at another time t
Σi ρi <Φi | E0 ? μ E0 ? μ (t) | Φi>
That is, these rates can be expressed either in a state-to-state manner or in a time-
dependent correlation function framework. In Chapter 7, this same correlation function
approach was examined further.
In an analogous fashion, it is possible to express chemical reaction rate constants
in a time-domain language again using time correlation functions. The TST (or VTST)
and RRKM expressions for the rate constant krate all involve, through the partition
26
functions or state densities, the reactant and transition-state energy levels and
degeneracies. These theories are therefore analogs of the state-to-state photon-absorption
rate equations.
To make the connection between the state-to-state and time-correlation function
expressions, one can begin with a classical expression for the rate constant given below:
Here Qr is the partition function of the reactant species, L is the number of coordinates
and momenta upon which the Hamiltonian H(p,q) depends, and β is 1/kT. The flux factor
F and the reaction probability χ are defined in terms of a dividing surface which could,
for example, be a plane perpendicular to the reaction coordinate S and located along
the reaction path that was discussed earlier in this Chapter in Section IC. Points on such
a surface can be defined by specifying one condition that the L coordinates {qj} must
obey, and we write this condition as
f(q) = 0.
Points lying where f(q) < 0 are classified as lying in the reactant region of coordinate
space, while those lying where f > 0 are in the product region. For example, if the
dividing surface is defined as mentioned above as being a plane perpendicular to the
reaction path, the function f can be written as:
k(T) = Q
r
?1
(2pih)
? L
dpdqe
? βH(q, p)
∫ F(p,q)χ(p,q)
27
f(q) = (S - S0).
Here, S is the reaction coordinate (which, of course, depends on all of the q variables)
and S0 is the value of S at the dividing surface. If the dividing surface is placed at the
transition state on the energy surface, S0 vanishes because the transition state is then, by
convention, the origin of the reaction coordinate.
So, now we see how the dividing surface can be defined, but how are the flux
F and probability χ constructed? The flux factor F is defined in terms of the dividing
surface function f(q) as follows:
F(p,q) = d h(f(q))/dt
= (dh/df) (df/dt)
=(dh/df) Σj ?f/?qj (dqj/dt)
= δ(f(q)) Σj ?f/?qj (dqj/dt).
Here, h(f(q)) is the Heaviside step function (h(x) = 1 if x>0; h(x) = 0 if x < 0), whose
derivative dh(x)/dx is the Dirac delta function δ(x), and the other identities follow by
using the chain rule. When the dividing surface is defined in terms of the reaction path
coordinate S as introduced earlier (i.e., f(q) = (S - S0)), the factor Σj ?f/?qj (dqj/dt)
contains only one term when the L coordinates {qj} are chosen, as in the reaction path
28
theory, to be the reaction coordinate S and L-1 coordinates {q’j} = q’ perpendicular to the
reaction path. For such a choice, one obtains
Σj ?f/?qj (dqj/dt) = dS/dt = PS/mS
where PS is the momentum along S and mS is the mass factor associated with S in the
reaction path Hamiltonian. So, in this case, the total flux factor F reduces to:
F(p,q) = δ(S-S0) PS/mS.
We have seen exactly this construct before in Section IA where the TST expression for
the rate coefficient was developed.
The reaction probability factor χ(p,q) is defined in terms of those trajectories that
evolve, at long time t → ∞, onto the product side of the dividing surface; such trajectories
obey
χ(p,q) = limt → ∞ h(f(q(t))) = 1.
This long-time limit can, in turn, be expressed in a form where the flux factor again
occurs:
lim
t →∞
h( f (q(t))) =
dh( f (q(t)))
dt
0
∞
∫ dt = Fdt
0
∞
∫
29
In this expression, the flux F(t) pertains to coordinates q(t) and momenta p(t) at t > 0.
Because of time reversibility, the integral can be extended to range from from t = - ∞ to t
= ∞.
Using the expressions for χ and for F as developed above in the equation for the
rate coefficient given at the beginning of this Section allows the rate coefficient k(T) to
be rewritten as follows:
In this form, the rate constant k(T) appears as an equilibrium average (represented by the
integral over the initial values of the variables p and q with the Qr-1 (2pih)-L exp(-βH)
weighting factor) of the time correlation function of the flux F:
To evaluate the rate constant in this time-domain framework for a specific
chemical reaction, one would proceed as follows.
i. Run an ensemble of trajectories whose initial coordinates and momenta {q.p} are
selected (e.g., using Monte-Carlo methods discussed in Chapter 7) from a distribution
with exp(-βH) as its weighting factor.
k(T) = Q
r
?1
(2pih)
? L
dpdqe
? βH(q, p)
∫ F(p,q)χ(p,q)
= Q
r
?1
(2pih)
? L
dt
?∞
∞
∫ dpdqe
?βH (q, p)
∫ F( p,q)F( p(t),q(t))
dt
?∞
∞
∫ F(p,q)F( p(t),q(t))
30
ii. Make sure that the initial coordinates {q} lie on the dividing surface because the flux
expression contains the δ(f(q)) factor;
iii. Monitor each trajectory to observe when it again crosses the dividing surface (i.e.,
when {q(t)} again obeys f(q(t)) = 0; at which time the quantity
iv. F(p(t),q(t)) can be evaluated as F(p,q) = δ(f(q)) Σj ?f/?qj (dqj/dt), using the coordinates
and momenta at time t to compute these quantities.
Using a planar dividing surface attached to the reaction path at S = S0 as noted
earlier allows F(q,p) to be calculated in terms of the initial (t=0) momentum lying along
the reaction path direction as , F(p,q) = δ(S-S0) PS/mS and permits F(p(t),q(t)) to be
computed when the trajectory again crosses this surface at at time t as F(p(t),q(t)) = δ(S-
S0) PS(t)/mS. So, all that is really needed if the dividing surface is defined in this manner
is to start trajectories with S = S0; to keep track of the initial momentum along S; to
determine at what times t the trajectory returns to S = S0; and to form the product (PS/mS)
(PS(t)/mS) for each such time. It is in this manner that one can compute flux-flux
correlation functions and, thus, the rate coefficient.
Notice that trajectories that undergo surface re-crossings contribute negative
terms to the flux-flux correlation function computed as discussed above. That is, a
trajectory with a positive initial value of (PS/mS) can, at some later time t, cross the
dividing surface with a negative value of (PS(t)/mS) (i.e, be directed back toward
reactants). This re-crossing will contribute a negative value, via. the product (PS/mS)
(PS(t)/mS), to the total correlation function, which integrates over all times. Of course, if
this same trajectory later undergoes yet another crossing of the dividing surface at t' with
31
positive PS(t'), it will contribute a positive term to the correlation function via. (PS/mS)
(PS(t')/mS). Thus, the correlation function approach to computing the rate coefficient can
properly account for surface re-crossings, unlike the TST which requires one to account
for such effects (and tunneling) in the transmission coefficient κ.
G. Wave Packet Propagation
The preceding discussion should have made it clear that it is very difficult to
propagate wave functions rigorously using quantum mechanics. On the other hand, to
propagate a classical trajectory is relatively straightforward. There exists a powerful tool
that allows one to retain much of the computational ease and convenient interpretation of
the classical trajectory approach while also incorporating quantum effects that are
appropriate under certain circumstances. In this wave packet propagation approach, one
begins with a quantum mechanical wave function that is characterized by two parameters
that give the average value of the position and momentum along each coordinate. One
then propagates not the quantum wave function but the values of these two parameters,
which one assumes will evolve according to Newtonian dynamics. Let's see how these
steps are taken in more detail and try to understand when such an approach is expected to
work or to fail.
First, the form of the so-called wave packet quantum function is written as
follows:
Ψ(q,Q, P) = ΠJ=1,N (2pi<δqJ2>)-1/2 exp[(iPJ qJ/ h) -(qJ - QJ)2 /4<δqJ2>].
32
Here, we have a total of N coordinates that we denote {qJ : J = 1, N}. It is these
coordinates that the quantum wave function depends upon. The total wave function is a
product of terms, one for each coordinate. Notice that this wave function has two distinct
ways in which the coordinate qJ appear. First, it has a Gaussian (exp[- (qJ - QJ)2 /4<δqJ2>])
dependence centered at the values QJ and having Gaussian width factors related to <qJ2>.
This dependence tends to make the wave function's amplitude largest when qJ is close to
QJ. Secondly, it has a form exp[(iPJ qJ/ h)] that looks like the travelling wave that we
encountered in the Background Material in which the coordinate qJ moves with
momentum PJ . So, these wave packet functions have built into them characteristics that
allow them to describe motion (via. the PJ) of an amplitude that is centered at QJ with a
width given by the parameter <qJ2>.
The parameters PJ and QJ we assume in this approach to chemical dynamics will
undergo classical time evolution according to the Newton equations:
dQJ/dt = PJ/mJ
dPJ/dt = - ?E/?QJ
where E is the potential energy surface (Born-Oppenheimer or force field) upon which
we wish to propagate the wave packet, and mJ is the mass associated with coordinate qJ .
The QJ and PJ parameters can be shown to be the expectation values of the coordinates qJ
and momenta -i h?/?qJ for the above function:
33
QJ = ∫ Ψ* qJ Ψ dq,
PJ = ∫ Ψ* (- i h ?/?qJ) Ψ dq.
Moreover, the <qJ2> parameter appearing in the Gaussian part of the function can be
shown to equal the dispersion or "spread" of this wave function along the coordinate qJ:
<qJ2> = ∫ Ψ* (qJ - QJ)2 Ψ dq.
There is an important characteristic of the above Gaussian wave packet functions
that we need to point out. It turns out that functions of this form:
Ψ(q,Q(t), P(t)) = ΠJ=1,N (2pi<δqJ2>)-1/2 exp[(iPJ(t) qJ/ h) -(qJ - QJ(t))2 /4<δqJ2>]
can be shown to have uncertainties in qJ and in - i h ?/?qJ that are as small as possible:
<(qJ –QJ)2> <(- i h ?/?qJ – PJ)2> = h2/4.
The Heisenberg uncertainty relation, which is discussed in many texts dealing with
quantum mechanics, says that this product of coordinate and momentum dispersions must
be greater than or equal to h2/4. In a sense, the Gaussian wave packet function is the most
34
classical function that one can have because its uncertainty product is as small as possible
(i.e, equals h2/4). We say this is the most classical possible quantum function because in
classical mechanics, both the coordinate and the momentum can be known precisely. So,
whatever quantum wave function allows these two variables to be least uncertain is the
most classical.
To use wave packet propagation to simulate a chemical dynamics event, one
begins with a set of initial classical coordinates and momenta {QJ(0), PJ(0)} as well as a
width <qJ2> or uncertainty for each coordinate. Each width must be chosen to represent
the range of that coordinate in the experiment that is to be simulated. For example,
assume one were to represent the dynamics of a wave function that is prepared by photon
absorption of a v = 0 vibrational state of the H-Cl molecule from the ground 1Σ state to an
excited-state energy surface (E(R)). Such a situation is described qualitatively in Fig. 8.3.
In this case, one could choose <δR2> to be the half width of the v = 0 harmonic (or
Morse) oscillator wave function χ0(R) of H-Cl, and take P(0) = 0 (because this is the
average value of the momentum for χ0) and R(0) = Req, the equilibrium bond length.
For such initial conditions, classical Newtonian dynamics would then be used to
propagate the QJ and PJ. In the H-Cl example, introduced above, this propagation would
be performed using the excited-state energy surface for E since, for t > 0, the molecule is
assumed to be on this surface. The total energy at which the initial wave packet it
delivered to the upper surface would be dictated by the energy of the photon used to
perform the excitation. In Fig. 8.3, two such examples are shown.
Once the packet is on the upper surface, its position Q and momentum P begin to
change according to the Newton equations. This, in turn, causes the packet to move as
35
shown for several equally spaced time “steps” in Fig. 8.3 for the two different photons’
cases. At such subsequent times, the quantum wave function is then assumed, within this
model, to be given by:
Ψ(q,Q(t), P(t)) = ΠJ=1,N (2pi<δqJ2>)-1/2 exp[(iPJ(t) qJ/ h) -(qJ - QJ(t))2 /4<δqJ2>].
That is, it is taken to be of the same form as the initial wave function but to have simply
moved its center from Q(0) to Q(t) with a momentum that has changed from P(0) to P(t).
36
Ground-State Potential
Excited-State Potential
Initital Wave Packet
Wave Packet Just After Absorption of Low-Energy
Photon
Wave Packet Just After Absorption of Higher-Energy
Photon
t=0 t1 t2
t3
t=0 t1 t2
t3 t4 t5
Figure 8.3 Propagation of Wave Packet Prepared by Absorption of Two Different
Photons.
It should be noticed that the time evolution of the wave packet shown in Fig. 8.3 displays
clear classical behavior. For example, as time evolves, it moves to large R-values and its
speed (as evidenced by the spacings between neighboring packets for equal time steps) is
37
large when the potential is low and small when the potential is higher. As we learned in
the Background Material and in Chapter 6, the time correlation function
C(t) = <Ψ(q,Q(0),P(0)) | Ψ(q,Q(t),P(t))>
can be used to extract spectral information by Fourier transformation. For the H-Cl
example considered here, this correlation function will be large at t = 0 but will decay in
magnitude as the wave packet Ψ(q,Q(t),P(t)) moves to the right (at t1, t2, etc.) because its
overlap with Ψ(q,Q(0),P(0)) becomes smaller and smaller as time evolves. This decay in
C(t) will occur more rapidly for the high-energy photon case because Ψ(q,Q(t),P(t))
moves to the right more quickly because the classical momentum P(t) grows more
rapidly. These dynamics will induce exponential decays in C(t) (i.e., C(t) will vary as
exp(-t/τ1)) at short times.
In fact, the decay of C(t) discussed above produces, when C(t) is Fourier
transformed, the primary characteristic of the correlation function for the higher-energy
photon case where dissociation ultimately occurs. In such photo-dissociation spectra, one
observes a Lorentzian line shape whose width is characterized by the decay rate (1/τ1),
which, in turn, relates to the total energy of the packet and the steepness of the excited-
state surface. This steepness determines how fast P(t) grows, which then determines how
fast the H-Cl bond fragments.
In the lower-energy photon case shown in Fig. 8.3, a qualitatively different
behavior occurs in C(t) and thus in the spectrum. The packet’s movement to larger R
causes C(t) to initially undergo exp(-t/τ1) decay. However, as the packet moves to its
38
large-R turning point (shortly after time t3), it strikes the outer wall of the surface where
it is reflected. Subsequently, it undergoes motion to smaller R, eventually returning to its
initial location. Such recurrences, which occur on time scales that we denote τ2, are
characteristic of bound motion in contrast to the directly dissociative motion discussed
earlier. This recurrence will cause C(t) to again achieve a large amplitude, but, C(t) will
subsequently again undergo exp(-t/τ1) decay as the packet once again departs. Clearly, the
correlation function will display a series of recurrences followed by exponential decays.
The frequency of the recurrences is determined by the frequency with which the packet
traverses from its inner to outer turning points and back again, which is proportional to
1/τ2. This, of course, is the vibrational period of the H-Cl bond. So, in such bound-motion
cases, the spectrum (i.e., the Fourier transform of C(t)) will display a series of peaks
spaced by (1/τ2) with the envelope of such peaks having a width determined by 1/τ1.
In more complicated multi-mode cases (e.g., in molecules containing several
coordinates), the periodic motion of the wave packet usually shows another feature that
we have not yet discussed. Let us, for simplicity, consider a case in which only two
coordinates are involved. For the wave packet to return to (or near) its initial location
enough time must pass for both coordinates to have undergone an excursion to their
turning points and back. For example, consider the situation in which one coordinate’s
vibrational frequency is ca. 1000 cm-1 and the other’s is 300 cm-1; these two modes then
require ca. 1/30 ps and 1/9 ps, respectively, to undergo one complete oscillation. At t = 0,
the wave packet, which is a product of two packets, ΠJ=1,2 (2pi<δqJ2>)-1/2 exp[(iPJ(t) qJ/ h) -
(qJ - QJ(t))2 /4<δqJ2>], one for each mode, produces a large C(t). After 1/30 ps, the first
mode’s coordinate has returned to its initial location, but the second mode is only 9/30 of
39
the way along in its periodic motion. Moreover, after 1/9 ps, the second mode’s
coordinate has returned to near where it began, but now the first mode has moved away.
So, at both 1/30 ps and 1/9 ps, the correlation function will not be large because one of
the mode contribution to C(t) = <Ψ(q,Q(0),P(0)) | Ψ(q,Q(t),P(t))> will be small.
However, after 1/3 ps, both modes’ coordinates will be in positions to produce a large
value of C(t); the high-frequency mode will have undergone 10 oscillations, and the
lower-frequency mode will have undergone 3 oscillations. My point in discussing this
example has been to illustrate that molecules having many coordinates can produce
spectra that display rather complicated patterns but which, in principle, can be related to
the time evolution of these coordinates using the correlation function’s connection to the
spectrum.
Of course, there are problems that arise in using the wave packet function to
describe the time evolution of a molecule (or any system that should be treated using
quantum mechanics). One of the most important limitations of the wave packet approach
to be aware of relates to it inability to properly treat wave reflections. It is well know that
when a wave strikes a hard wall it is reflected by the wall. However, when, for example,
a water wave moves suddenly from a region of deep water to a much more shallow
region, one observes both a reflected and a transmitted wave. In the discussion of
tunneling resonances given in the Background Material, we also encountered reflected
and transmitted waves. Furthermore, when a wave strikes a barrier that has two or more
holes or openings in it, one observes wave fronts coming out of these openings. The
problem with the most elementary form of wave packets presented above is that each
packet contains only one “piece”. It therefore can not break into two or more “pieces” as
40
it, for example, reflects from turning points or passes through barriers with holes.
Because such wave packets can not fragment into two or more packets that subsequently
undergo independent dynamical evolution, they are not able to describe dynamical
processes that require multiple-fragmentation events. It is primarily for this reason that
wave packet approaches to simulating dynamics are usually restricted to treating short-
time dynamics where such fragmentation of the wave packet is less likely to occur.
Prompt molecular photo-dissociation processes such as we discussed above is a good
example of such a short-time phenomenon.
H. Surface Hopping Dynamics
There are, of course, chemical reactions and energy transfer collisions in which
two or more Born-Oppenheimer energy surfaces are involved. Under such circumstances,
it is essential to have available the tools needed to describe the coupled electronic and
nuclear-motion dynamics appropriate to this situation.
The way this problem is addressed is by returning to the Schr?dinger equation
before the single-surface Born-Oppenheimer approximation was made and expressing the
electronic wave function Ψ(r|R), which depends on the electronic coordinates {r} and
the nuclear coordinates {R}, as:
Ψ(r|R) = ΣJ aJ(t) ψJ(r|R) exp[-i/ h ∫ HJ,J (R) dt].
41
Here, ψJ(r|R) is the Born-Oppenheimer electronic wave function belonging to the Jth
electronic state, and HJ,J(R) is the expectation value of H for this state. The aJ(t) are
amplitudes that will eventually relate to the probability that the system is “on” the Jth
surface. Substituting this expansion into the time-dependent Schr?dinger equation
i h ?Ψ/?t = H Ψ
followed by multiplying on the left by ψ*K(r|R) and integrating over the electronic
coordinates {r}, gives an equation for the aK(t) amplitudes:
i h daK/dt = ΣJ≠K { HK,J - i h <ψK|dψJ/dt>} exp[-i/ h ∫ (HJ,J (R) – HK,K(R)) dt].
Here, HK,J is the Hamiltonian matrix that couples ψK to ψJ. The integral appearing in the
exponential is taken from an initial time ti when one is assumed to know that the system
resides on a particular Born-Oppenheimer surface (K) up to the time t, at which the
amplitude for remaining on this surface aK as well as the amplitudes for hopping to other
surfaces {aJ} are needed. This differential equation for the amplitudes is solved
numerically by starting at ti with aK = 1 and aJ≠K = 0 and propagating the amplitudes’
values forward in time.
The next step is to express <ψK|dψJ/dt>, using the chain rule, in terms of
derivatives with respect to the nuclear coordinates {R} and the time rate of change of
these coordinates:
42
<ψK|dψJ/dt> = Σa <ψK|dψJ/dRa> dRa/dt.
This is how a classical dynamical treatment of the nuclear motions will be introduced
(i.e., by assuming that the nuclear coordinates Ra and dRa/dt obey Newton equations). So,
now the equations for the aK(t) read as follows:
i h daK/dt = ΣJ≠K { HK,J - i h Σa dRa/dt <ψK|dψJ/dRa>} exp[-i/ h ∫ (HJ,J (R) – HK,K(R)) dt].
The <ψK|dψJ/dRa> are called non-adiabatic coupling matrix elements, and it is their
magnitudes that play a central role in determining how efficient surface hoppings are.
These matrix elements are becoming more commonly available in widely utilized
quantum chemistry and dynamics computer packages (although their efficient evaluation
remains a challenge that is undergoing significant study).
In addition to the above prescription for calculating amplitudes (the probabilities
then being computed as |aJ|2), one also needs to identify (using, perhaps the kind of
strategy discussed in Chapter 3) the seam at which the surfaces of interest intersect. Let
us, for example, assume that there are two surfaces that undergo such an intersection in a
subspace of the 3N-6 dimensional energy surfaces H1,1 and H2,2, and let us denote nuclear
coordinates {Ra} that lie on this seam as obeying
F(Ra) = 0.
following the discussion of Chapter 3.
43
To utilize the most basic form of surface hopping theory, one proceeds as follows:
1. One begins with initial values of the nuclear coordinates {Ra} and their velocities
{dRa/dt} that properly characterize the kind of collision or reaction one wishes to
simulate. Of course, one may have to perform many such surface hopping trajectories
with an ensemble of initial conditions chosen to properly describe such an experimental
situation. In addition, one specifies which electronic surface (say the Kth surface) the
system is initially on.
2. For each such set of initial conditions, one propagates a classical trajectory describing
the time evolution of the {Ra} and {dRa/dt} on this initial surface, until such a trajectory
approaches a geometry lying on the intersection seam.
3. As one is propagating the classical trajectory from ti up to the time t1 when it
approaches the seam, one also propagates the (in this example, two) differential equations
i h daK/dt = ΣJ≠K { HK,J - i h Σa dRa/dt <ψK|dψJ/dRa>} exp[-i/ h ∫ (HJ,J (R) – HK,K(R)) dt]
that produce the time evolution of the {aJ} amplitudes.
4. Once the classical trajectory reaches the seam, the initial trajectory is halted and two
(in the case of M coupled surfaces, M) new trajectories are initiated, one beginning (at t1)
on each of the (two in our case) coupled surfaces. Each of these new trajectories (two )
are assigned a probability weighting factor computed as the square of the amplitude
belonging to the surface upon which the trajectory now resides: |aJ|2.
5. Each of the new trajectories is subsequently allowed to propagate from t1 (where the
nuclear coordinates are what they were just prior to diving the one initial trajectory into
two (or M) trajectories) but now on different Born-Oppenheimer surfaces. This
44
propagation is continued until one of these trajectories again approaches a geometry on
the intersection seam, at which time the trajectory hopping scheme is repeated.
Clearly, it is possible that the one initial trajectory will generate a large number of
subsequent trajectories each of which is assigned an overall weighting factor taken to be
the product of all of the |aJ|2 probabilities associated with trajectories from which this
particular trajectory was descended. It is these overall weightings that one uses to predict,
one this process has been carried out long enough to allow the reaction or collision to
proceed to completion, the final yield of products in each Born-Oppenheimer state. This
surface-hopping algorithm remains one of the most widely used approaches to treating
such coupled-state dynamics.
II. Experimental Probes of Reaction Dynamics
A. Spectroscopic Methods
To follow the rate of any chemical reaction, one must have a means of monitoring
the concentrations of reactant or product molecules as time evolves. In the majority of
current experiments that relate to reaction dynamics, one uses some form of
spectroscopic or alternative physical probe (e.g., an electrochemical signature) to monitor
45
these concentrations as functions of time. Of course, in all such measurements, one must
know how the intensity of the signal detected relates to the concentration of the
molecules that cause the signal. For example, in most absorption experiments, as
illustrated in Fig. 8.4, light is passed through a sample of “thickness” L and the intensity
of the light beam in the absence of the sample I0 and with the sample present I are
measured.
Figure 8.4 Typical Beer’s –Law Experiment in Which a Light Beam of Intensity I0 is
Passed Through a Sample of Thickness L.
The Beer-Lambert law:
log(I0/I) = ε [A] L
Intensity ofIncident Light I
0
Intensity ofTransmitted Light I
Sample of thickness L
46
then allows the concentration [A] of the absorbing molecules to be determined, given the
path length L over which absorption occurs and given the extinction coefficient ε of the
absorbing molecules.
These extinction coefficients, which relate to the electric dipole matrix elements
that are discussed in Chapter 6, are usually determined empirically by preparing a known
concentration of the absorbing molecules and measuring the I0/I ratio that this
concentration produces in a cell of length L. For molecules and ions that are extremely
reactive, this “calibration” approach to determining ε is often not feasible because one
can not prepare a sample with a known and stable concentration. In such cases, one often
must resort to using the theoretical expressions given in Chapter 6 (and discussed in most
textbooks on molecular spectroscopy) to compute ε in terms of the wave functions of the
absorbing species. In any event, one must know how the strength of the signal relates to
the concentrations of the species if one wishes to monitor chemical reaction or energy
transfer rates.
Because modern experimental techniques are capable of detecting molecules in
particular electronic and vibration-rotation states, it has become common to use such
tools to examine chemical reaction dynamics on a state-to-state level and to follow
energy transfer processes, which clearly require such state-specific data. In such
experiments, one seeks to learn the rate at which reactants in a specific state Φi react to
produce products in some specific state Φf. One of the most common ways to monitor
such state-specific rates is through a so-called “pump-probe” experiment in which
i. A short-duration light source is used to excite reactant molecules to some
specified initial state Φi. Usually a tunable laser is used because its narrow
47
frequency spread allows specific states to be pumped. The time at which this
“pump” laser thus prepares the excited reactant molecules in state Φι defines t = 0.
ii. After a “delay time” of duration τ, a second light source is used to “probe” the
product molecules that have been formed in various final states, Φf. Usually, the
frequency of this probe source is scanned so that one can examine populations of
many such final states.
The concentrations of reactant and products molecules in the initial and final states, Φi
and Φf, are determined by the Beer-Lambert relation assuming that the extinction
coefficients εi and εf for these species and states absorption are known. In the former case,
the extinction coefficient ει relates to absorption of the pump photons to prepare reactant
molecules in the specified initial state. In the latter, εf refers to absorption of the product
molecules that are created in the state Φf. Carrying out a series of such final-state
absorption measurements at various delay times τ allows one to determine the
concentration of these states as a function of time.
This kind of laser pump-probe experiment is used not only to probe specific
electronic or vibration/rotation states of the reactants and products but also when the
reaction is fast (i.e., complete in 10-4s or less). In these cases, one is not using the high
frequency resolution of the laser but its fast time response. Because laser pulses of quite
short duration can be generated, these tools are well suited in such fast chemical reaction
studies. The reactions can be in the gas phase (e.g., fast radical reactions in the
atmosphere or in explosions) or in solution (e.g., photo-induced electron transfer
reactions in biological systems).
48
B. Beam Methods
Another approach to probing chemical reaction dynamics is to use a beam of reactant
molecules A that collides with other reactants B that may also in a beam or in a “bulb” in
equilibrium at some temperature T. Such crossed-beam and beam-bulb experiments are
illustrated in Fig. 8.5.
49
Figure 8.5 Typical Crossed-Beam and Beam-Bulb Experimental Setups.
Almost always, these beam and bulb samples contain molecules, radicals, or ions in the
gas phase, so these techniques are most prevalent in gas-phase dynamics studies.
The advantages of the crossed-beam type experiments are that:
(i) one can control the velocities, and hence the collision energies, of both reagents,
Beam of A molecules
moving with velocity
v.
Product Cmolecules
detected exitingwith velocity v'
in internal state jat angle θθ
Bulb containingB molecules at
temperature T
Beam of A molecules
moving with velocity
v.
Product Cmolecules
detected exitingwith velocity v''
in internal state jat angle θθ
Beam of B molecules
moving with velocity
v'.
Collision chamberin which A and B
collide and react
50
(ii) one can examine the product yield as a function of the angle θ through which the
products are scattered,
(iii) one can probe the velocity of the products and,
(iv) by using spectroscopic methods, one can determine the fraction of products
generated in various internal (electronic/vibrational/rotational) states.
Such measurements allow one to gain very detailed information about how the
reaction rate coefficient depends on collisional (kinetic) energy and where the total
energy available to the products is deposited (i.e., into product translational energy or
product internal energy). The angular distribution of product molecules can also give
information about the nature of the reaction process. For example, if the A + B collision
forms a long-lived (i.e., on rotational time scales) collision complex, the product C
molecules display a very isotropic angular distribution.
In beam-bulb experiments, one is not able to gain as much detailed information
because one of the reactant molecules B is not constrained to be moving with a known
fixed velocity in a specified direction when the A + B → C collisions occur. Instead, the
B molecules collide with A molecules in a variety of orientations and with a distribution
of collision energies whose range depends on the Maxwell-Boltzmann distribution of
kinetic energies of the B molecules in the bulb. The advantage of beam-bulb experiments
is that one can achieve much higher collision densities than in crossed-beam experiments
because the density of B molecules inside the bulb is much higher than are the densities
achievable in a beam of B molecules.
There are cases in which the beam-bulb experiments can be used to determine how
the reaction rate depends on collision energy even though the molecules in the bulb have
51
a distribution of kinetic energies. That is, if the species in the beam have much higher
kinetic energies than most of the B molecules, then the A + B collision energy is
primarily determined by the beam energy. An example of this situation is provided by so-
called guided-ion beam experiments in which a beam of ions having well-specified
kinetic energy E impinges on molecules in a bulb having a temperature T for which kT
<<E. Fig. 8.6 illustrates data that can be extracted from such an experiment.
52
Figure 8.6 Collision-Induced Dissociation Data Showing Cross-Section as a Function of
Collision Energy.
In Fig. 8.6, we illustrate the cross-section σ (related to the bimolecular rate constant k by
σ v = k, where v is the relative collision speed) for production of Na+ ions when a beam
of Na+(uracil) complexes having energy E (the horizontal axis) collides with a bulb
containing Xe atoms at room temperature. In this case, the reaction is simply the
collision-induced dissociation (CID) process in which the complex undergoes
unimolecular decomposition after gaining internal energy in its collsions with Xe atoms:
Na+(uracil) → Na+ + uracil.
The primary knowledge gained in this CID experiment is the threshold energy E*; that is,
the minimum collision energy needed to effect dissociation of the Na+(uracil) complex.
This kind of data has proven to offer some of the most useful information about bond
dissociation energies of a wide variety of species. In addition, the magnitude of the
reaction cross-section σ as a function of collision energy is a valuable product of such
experiments. These kind of CID beam-bulb experiments offer one of the most powerful
and widely used means of determining such bond-rupture energies and reaction rate
constants.
C. Other Methods
53
Of course, not all chemical reactions occur so quickly that they require the use of
fast lasers to follow concentrations of reacting species or pump-probe techniques to
generate and probe these molecules. For slower chemical reactions, one can use other
methods for monitoring the relevant concentrations. These methods include
electrochemistry (where the redox potential is the specie's signature) and NMR
spectroscopy (where the chemical shifts of substituents are the signatures) both of whose
instrumental response times would be too slow for probing fast reactions.
In addition, when the reactions under study do not proceed to completion but exist in
equilibrium with a back reaction, alternative approaches can be used. The example
discussed in Chapter 5 is one such case. Let us briefly review it here and again consider
the reaction of an enzyme E and a substrate S to form the enzyme-substrate complex ES:
E + S ? ES.
In the perturbation-type experiments, the equilibrium concentrations of the species are
"shifted" by a small amount δ by application of the perturbation, so that
[ES] = [ES]eq -δ
[E] = [E]eq + δ
[S] = [S]eq + δ.
54
Subsequently, the following rate law will govern the time evolution of the concentration
change δ:
- dδ/dt = - kr ([ES]eq -δ) + kf ([E]eq + δ) ([S]eq + δ).
Assuming that δ is very small (so that the term involving δ2 cam be neglected) and using
the fact that the forward and reverse rates balance at equilibrium, this equation for the
time evolution of δ can be reduced to:
- dδ/dt = (kr + kf [S]eq + kf [Eeq]) δ.
So, the concentration deviations from equilibrium will return to equilibrium
exponentially with an effective rate coefficient that is equal to a sum of terms:
keff = kr + kf [S]eq + kf [Eeq].
So, by following the concentrations of the reactants or products as they return to
their equilibrium values, one can extract the effective rate coefficient keff. Doing this at a
variety of different initial equilibrium concentrations (e,g., [S]eq and [E]eq), and seeing
how keff changes, one can then determine both the forward and reverse rate constants.