Computational Chemistry Course Overview
Computational Chemistry Course Overview
+ =
There are two different aspects to computational chemistry:
1. Computational studies can be carried out to find a starting point for a laboratory synthesis, or to
assist in understanding experimental data, such as the position and source of spectroscopic peaks.
2. Computational studies can be used to predict the possibility of so far entirely unknown molecules
or to explore reaction mechanisms that are not readily studied by experimental means.
Thus, computational chemistry can assist the experimental chemist or it can challenge the experimental
chemist to find entirely new chemical objects.
Importance of computational chemistry
Computational methods are considered important in scientific community, see:
The top 100 papers in science
Also, computational chemistry has featured in a number of Nobel Prize awards, most notably in 1998
and 2013.
Some applications
Drug design
Atmospheric science
Computational chemistry programs
There are many self-sufficient software packages used by computational chemists. Some include
many methods covering a wide range, while others concentrating on a very specific range or even
a single method. Details of most of them can be found in:
•Biomolecular modelling programs: proteins, nucleic acid.
•Molecular mechanics programs.
•Quantum chemistry and solid state physics software supporting several methods.
•Molecular design software
•Semi-empirical programs.
•Valence bond programs.
Challenges and limitations
• Theory level: Is the precision high enough for the application?
• The computer programs: Efficiency and implementation
• Computer resources: Is there enough power for the task?
About CSC
CSC IT Center for Science Ltd is administered by the
Ministry of Education, Science and Culture.
CSC maintains and develops the state-owned centralised IT
infrastructure and uses it to provide nationwide IT services
for research, libraries, archives, museums and culture as
well as information, education and research management.
Researchers can use it’s large collection of scientific
software and databases.
CSC has offices in Espoo's Keilaniemi and in the Renforsin
Ranta business park in Kajaani.
Methods of computational chemistry
• Ab initio that use rigorous quantum mechanics
+ accurate
– computationally expensive
good results for small systems ~102 atoms
• Semi empirical that use approximate quantum mechanics
– relies on empirical or ab initio parameters
+ affordable and in some cases accurate
limited to well defined systems with ~104 atoms
• Molecular mechanics that use classical mechanics
– relies on empirical force fields without accounting for
electronic properties (no bond breaking or forming)
+ very affordable and used as a virtual experiment
can handle extremely large systems ~109 atoms
Principles of quantum chemistry
• The state of the system is specified by a (normalized) wave function 𝜓
• For every measurable property (observable) of a system such as energy 𝐸 for example, there exist
a corresponding operator (𝐻 for 𝐸)
• Observables satisfy the eigenvalue equation. For example 𝐻𝜓 = 𝐸𝜓.
• The expectation value of the observable, for example 𝐸, is given by
∞
𝐸= 𝜓 ∗ 𝐻𝜓 𝑑𝜏
−∞
𝑐0
𝑐1
For hydrogen atom, the optimal solution is obviously found when 𝑐0 = 1 and 𝑐1 = 0.
In reality we don’t know the eigenfunctions. Instead we use basis functions which are physically relevant for
the problem in hand.
𝜙 = 𝑐0 𝜑0 + 𝑐1 𝜑1
The hydrogenic energy corresponding to this trial function is
𝑐02 𝜑0 𝐻ℎ 𝜑0 + 𝑐12 𝜑1 𝐻ℎ 𝜑1 + 𝑐0 𝑐1 𝜑0 𝐻ℎ 𝜑1 + 𝑐1 𝑐0 𝜑1 𝐻ℎ 𝜑0
𝐸𝜙 =
𝑐02 + 𝑐12
𝜕𝐸
= 0 → 𝜑0 𝐻ℎ 𝜑1 𝑐0 + ( 𝜑1 𝐻ℎ 𝜑1 − 𝐸)𝑐1 = 0
𝜕𝑐1
We have a system of linear equations. According to linear algebra the energy eigenvalues can be
obtained by finding solutions of a characteristic equation
𝜑0 𝐻ℎ 𝜑0 − 𝐸 𝜑0 𝐻ℎ 𝜑1 𝐻0,0 − 𝐸 𝐻0,1
= =0
𝜑0 𝐻ℎ 𝜑1 𝜑1 𝐻ℎ 𝜑1 − 𝐸 𝐻0,1 𝐻1,1 − 𝐸
The energies can be obtained by diagonalizing the Hamiltonian matrix
𝜑0 𝐻ℎ 𝜑0 𝜑0 𝐻ℎ 𝜑1 𝐻0,0 𝐻0,1 𝑑𝑖𝑎𝑔 𝐸0 0
𝐻= =
𝜑0 𝐻ℎ 𝜑1 𝜑1 𝐻ℎ 𝜑1 𝐻0,1 𝐻1,1 0 𝐸1
The secular determinant for N basis functions gives an N-th order polynomial in which is solved for N
different roots, each of which approximates a different eigenvalue.
Example: For a helium atom we can choose the trial function as follows:
𝜙 𝑟1 , 𝑟2 = 𝐶 1 + 𝑝𝑟12 + 𝑞 𝑟1 − 𝑟2 2 exp −𝛼 𝑟1 + 𝑟2
where 𝐶 is normalization constant and 𝑝, 𝑞, and 𝛼 are variational parameters.
After optimization:
𝑝 = 0.30, 𝑞 = 0.13, and 𝛼 = 1.816
𝐸 = −2.9024 a.u. (Three parameters)
𝐸 = −2.9037462 a.u. (1024 parameters)
𝐸 = −2.9037843 a.u. (Experimental value)
2 𝜕2 𝜕2 𝜕2
where ∇𝑖 = 2 + + , 𝑛 and 𝑁 are numbers, and 𝑚e and 𝑚n are masses of electrons and nuclei,
𝜕𝑥𝑖 𝜕𝑦𝑖2 𝜕𝑧𝑖2
respectively. The potential energy operator includes the electron-electron, nuclei-nuclei, and electron-nuclei
parts
𝑛 𝑛 𝑁 𝑁 𝑛 𝑁
1 𝑒2 1 𝑍𝑗 𝑍𝑗 ′ 𝑒 2 𝑍𝑗 𝑒 2
𝑉= + −
2 4𝜋𝜀0 𝐫𝑖 − 𝐫𝑖 ′ 2 4𝜋𝜀0 𝐑𝑗 − 𝐑𝑗 ′ 4𝜋𝜀0 𝐫𝑖 − 𝐑𝑗
𝑖 𝑖 ′ ≠𝑖 𝑗 𝑗 ′ ≠𝑗 𝑖 𝑗
Where 𝐫𝑖 and 𝐑𝑗 are positions of electrons and nuclei and 𝑍𝑗 is the atomic number of nuclei 𝑗. In practice, it is
impossible to solve the Scrödinger equation for the total wavefunction Ψ(𝐫𝑖 , 𝐑𝑗 ) exactly.
Practical solution: Let’s approximate the wavefunction in a form, where it is factorized in electronic
motion and nuclear motion parts
𝜓(𝐫𝑖 , 𝐑𝑗 ) ≈ 𝜓el (𝐫𝑖 ; 𝐑𝑗 )𝜓n (𝐑𝑗 )
where function 𝜓(𝐫𝑖 ; 𝐑𝑗 ) describes electronic motion (depending parametrically on the positions of
nuclei) and function 𝜓(𝐑𝑗 ) describes the nuclear motions (vibrations and rotations). With this
assumptions, the problem can be reformulated to two separate Scrödinger equations:
𝐻el 𝜓el 𝐫𝑖 ; 𝐑𝑗 = 𝑉(𝐑𝑗 )𝜓el 𝐫𝑖 ; 𝐑𝑗
𝐻n 𝜓n 𝐑𝑗 = 𝐸n 𝜓n 𝐑𝑗
The former equation is for the electronic problem, considering the nuclei to be fixed. The eigenvalue
𝑉(𝐑𝑗 ) can be called as interatomic potential, which is then used as a potential energy for the latter
equation for the nuclear motion problem.
This procedure, the so called Born-Oppenheimer approximation, is justified because electron is lighter
than the proton by the factor 2000, the electron quickly rearranges in response to the slower motion of
the nuclei.
Example: For the H2+ -ion the total energy operator can be written as
Using the Born-Oppenheimer approximation, we can write the electronic Hamiltonian (further
simplified by using atomic units a.u. or Hartree) as
∇2 1 1
𝐻el = − − −
2 𝑟1 𝑟2
The ground state trial wave function is
𝜓 = 𝑐 𝜓1𝑠1 + 𝜓1𝑠2
where the 1𝑠 functions are the type
3 2
1 1
𝜓1𝑠𝑖 = 𝑒 −𝑟𝑖 𝑎0
𝜋 𝑎0
Where constant 𝑎0 is the Bohr radius. The
normalization is
1
𝑐=
2 + 2𝑆12
Where the 𝑆12 is the overlap integral
between 𝜓1𝑠1 and 𝜓1𝑠2 functions
Note that because 𝐑 = 𝐫1 − 𝐫2 , the overlap integral and thus the ground state electronic energy depends
parametrically on the distance between the two nuclei 𝑅
𝑉(𝑅)
𝑅/bohr
The BO–approximation is justified when the energy gap
between ground and excited electronic states is larger than the
energy scale of the nuclear motion.
The BO–approximation breaks down when
• for example in metals, some semiconductors and graphene
the band gab is zero leading to coupling between electronic
motion and lattice vibrations (electron-phonon interaction)
• electronic transitions becomes allowed by vibronic coupling
(Herzberg-Teller effect)
The Jahn–Teller effect is responsible
• ground state degeneracies are removed by lowering the for the tetragonal distortion of the
symmetry in non-linear molecules (Jahn-Teller effect) hexaaquacopper(II) complex ion,
• interaction of electronic and vibrational angular momenta [Cu(OH2)6]2+, which might otherwise
in linear molecules (Renner-Teller effect) possess octahedral geometry. The two
axial Cu−O distances are 238 pm,
whereas the four equatorial Cu−O
distances are ~195 pm.
In the vicinity of conical intersections, the Born–Oppenheimer approximation breaks down,
allowing non-adiabatic processes to take place. The location and characterization of conical
intersections are therefore essential to the understanding to a wide range of reactions, such as
photo-reactions, explosion and combustion reactions, etc.
Potential energy surfaces
The potential energy surface (PES) is a central
concept in computational chemistry. A PES is a
relationship between energy of a molecular system
and its geometry. The BO-approximation makes the
concept of molecular geometry meaningful, makes
possible the concept of PES, and simplifies the
application of the Scrödinger equation to molecules.
Since the atoms are motionless while we hold the molecule at any given
geometry, its energy is not kinetic and it is by default potential
(“depending on position”). The geometry of the molecule is defined
using appropriate coordinate system (cartesian coordinates, internal
coordinates, Jacobi coordinates, etc.). The so called reaction coordinate,
which is important in describing the energy profile of a chemical
reaction, is simply some combination of bond distances and angles.
Among the main tasks of computational chemistry are to
determine the structure and energy of molecule and of the
transition states (TS) involved in chemical reactions.
The positions of the energy minima along the reaction
coordinate give the equilibrium structures of the reactants
and products. Similarly, the position of the energy
maximum defines the transition state.
Reactants, Products, and transition states are all stationary
points on the potential energy surface. This means that for
system with N atoms all partial derivatives of the energy
respect to each of the 3𝑁 − 6 independent geometrical
coordinates (𝑅𝑖 ) are zero:
𝜕𝑉
=0 𝑖 = 1,2, … , 3𝑁 − 6
𝜕𝑅𝑖
In the one-dimensional case, or along the reaction coordinate, reactants and products are located in
the energy minima and are characterized by a positive second energy derivative
𝑑2 𝑉
>0
𝑑𝑅2
The transition state is characterized by a negative second energy derivative
𝑑2 𝑉
<0
𝑑𝑅2
In the many-dimensional case, each independent geometrical coordinate, 𝑅𝑖 , gives rise to 3𝑁 − 6
𝜕2 𝑉 𝜕2 𝑉 𝜕2 𝑉
second derivatives: , ,…, . Thus, it is not possible to say whether any given
𝜕𝑅𝑖 𝑅1 𝜕𝑅𝑖 𝑅2 𝜕𝑅𝑖 𝑅3𝑁−6
coordinate corresponds to a stationary point. To see the correspondence, a new set of coordinates 𝜉𝑖 ,
referred as normal coordinates, is used. They have the property that their cross terms or non-diagonal
𝜕2 𝑉
terms in second energy derivative matrix vanish: = 0, etc. For the energy minima:
𝜕𝜉1 𝜉2
𝜕2𝑉
>0 𝑖 = 1,2, … , 3𝑁 − 6
𝜕𝜉𝑖2
Stationary points for which all but one of the second derivatives positive are so-called saddle
points and may correspond to transition states. If they do, the normal coordinate for which the
second derivative is negative is referred to as the reaction coordinate 𝜉𝑟 :
𝜕2𝑉
<0
𝜕𝜉𝑟2
Geometry optimization is the process of starting with an input
structure ”quess” and finding a stationary point on the PES. It’s
usually checked whether the stationary point is a minimum or a
transition state by calculating its vibrational frequencies. In
transition state one of the vibrations will possess negative (or
imaginary) harmonic frequency.
1 𝑘
𝜈=
2𝜋 𝜇
Where 𝑞𝑖 are displacements in some appropriate coordinate system (for example internal coordinate
displacements). We can find a new set of coordinates (Wilson’s GF method) that simplify the above
equation to the form:
𝑁
1 𝜕2𝑉 2
𝑉 𝜉1 , 𝜉2 , … , 𝜉𝑁 = 𝜉
2 𝜕𝜉𝑖2 𝑖
𝑖=1
Where 𝜉𝑖 are the normal coordinates of the molecule. All normal modes are independent in the
harmonic approximation.
Vibrations of a methylene group (-CH2-) in a molecule for illustration
The consequence of the Pauli principle here is that electrons of the same spin are kept apart by a repulsive
exchange interaction, which is a short-range effect, acting simultaneously with the long-range electrostatic or
Coulombic force. This effect is partly responsible for the everyday observation in the macroscopic world that
two solid objects cannot be in the same place at the same time.
Hartree-Fock theory
• In computational chemistry, the Hartree-Fock method has central importance
• If we have the HF solution, the accuracy can systematically be improved by
applying various techniques
• It is based on variational approach
• HF is an approximate method, close in spirit to the mean-field approach widely used
in solid state and statistical physics
∇2𝑛 𝑍𝐴 1 1
− − + 𝜓 = 𝐸𝜓
2 𝑅𝐴𝑛 2 𝑟𝑛𝑚
𝑛 𝐴 𝑛 𝑛 𝑚≠𝑛
When wavefunction 𝜓 is expressed as a single slater determinant (we approximate the real wavefunction
using products of single electron wavefunctions) of doubly occupied spatial orbitals we can write the
energy as
0
𝐸=2 𝜖𝑖 + 2𝐽𝑖𝑗 − 𝐾𝑖𝑗
𝑖 𝑖 𝑗
0
where the first zeroth-order energy term 𝜖𝑖 is the energy without electron-electron repulsion, 𝐽𝑖𝑗 is the
Coulomb integral and 𝐾𝑖𝑗 is the exchange integral
1
𝐽𝑖𝑗 = 𝜓𝑖∗ 1 𝜓𝑗∗ 2 𝜓 2 𝜓𝑖 1 𝑑𝜏1 𝑑𝜏2
𝑟12 𝑗
1
𝐾𝑖𝑗 = 𝜓𝑖∗ 1 𝜓𝑗∗ 2 𝜓 2 𝜓𝑗 1 𝑑𝜏1 𝑑𝜏2
𝑟12 𝑖
We now wish to find the orbitals which lead to a minimum value of the energy. The treatment is simplified
if we define coulomb and exchange operators as follows
1
𝐽𝑗 𝜓𝑖 1 = 𝜓𝑗∗ 2 𝜓 2 𝑑𝜏2 𝜓𝑖 1
𝑟12 𝑗
1
𝐾𝑗 𝜓𝑖 1 = 𝜓𝑗∗ 2 𝜓 2 𝑑𝜏2 𝜓𝑗 1
𝑟12 𝑖
Note that the exchange operator 𝐾𝑗 exchanges electron 1 and 2 between the two orbitals 𝜓𝑖 and 𝜓𝑗 . The
coulomb and exchange integrals can be written as
0
ℎ1 + 2𝐽𝑗 − 𝐾𝑗 𝜓 𝑖 1 = 𝐹 1 𝜓 𝑖 1 = 𝜖𝑖 𝜓 𝑖 1
𝑗
0 0 0
Where ℎ1 ℎ1 𝜓𝑖 1 = 𝜖𝑖 𝜓𝑖 1 is the ”hydrogen-like” Hamiltonian operator for electron 1 in the
field of bare nucleus (or nuclei), 𝐹 1 is Fock operator and 𝜖𝑖 is the orbital energy.
These equations show very clearly that in order to solve the one-electron orbital 𝜓𝑖 1 , it is necessary to
know wavefunctions 𝜓𝑗 in order to set up the operators 𝐽𝑗 and 𝐾𝑗 . The orbitals 𝜓𝑖 1 only account for
the presence of other electrons in an average manner (mean-field theory).
The Hartree–Fock method is also called the self-consistent field method (SCF), meaning that the final
field as computed from the charge distribution is required to be "self-consistent" with the assumed initial
field.
When the Hartree-Fock equations are solved by numerical integration methods, the procedure in unwieldy.
Even worse, the method is incapable of being extended to molecules (molecular orbital theory). The key
development (presented by Roothan) was to expand the orbitals 𝜓𝑖 as a linear combination of a set of one-
electron basis functions. Introducing a basis set transforms the Hartree-Fock equations into the Roothaan
equations. Denoting the atomic orbital basis functions as 𝜙𝑘 , we have the expansion
𝜓𝑖 = 𝑐𝑘𝑖 𝜙𝑘
𝑘
This leads to
𝐹 𝑐𝑘𝑖 𝜙𝑘 = 𝜖𝑖 𝑐𝑘𝑖 𝜙𝑘
𝑘 𝑘
1
𝐽𝑗 𝜙𝑙 1 = 𝜓𝑗∗ 2 𝜓 2 𝑑𝜏2 𝜙𝑙 1
𝑟12 𝑗
1
𝐾𝑗 𝜙𝑙 1 = 𝜓𝑗∗ 2 𝜙 2 𝑑𝜏2 𝜓𝑗 1
𝑟12 𝑙
Note that the coulomb and exchange operators depend on our initial guess.
The initial estimate of secular equation det 𝐹𝑘𝑙 − 𝜖𝑖 𝑆𝑘𝑙 = 0 is
−0.813 − 𝜖𝑖 −0.892 − 0.8366𝜖𝑖
=0
−0.892 − 0.8366𝜖𝑖 −0.070 − 𝜖𝑖
𝜖1 = −0.854
𝜖2 = 2.885
Substitution of the 𝜖1 into the Roothan equation gives improved orbital
𝜓𝑖 = 0.836𝜙1 + 0.189𝜙2
After few cycles the converged roots from secular equation are 𝜖1 = −0.918 and 𝜖2 = 2.810 and the final
orbital is 𝜓𝑖 = 0.842𝜙1 + 0.183𝜙2. To obtain the ground state energy of the system one needs to add to
the orbital energy (lower energy root 𝜖1 ) additional core electron energy term and in the case of molecules,
a nuclear-nuclear repulsion potential energy. So the SCF energy for helium becomes
𝐸HF = −0.918 − 1.944 + 0 = −2.862 hartrees
The limiting HF energy found with five basis functions is only 0.0000073 hartrees lower in energy. The
comparison to variational and experimental energies shows an approximate error of 0.042 hartrees (110
kJ/mol), which arises due to lack of electron correlation.
• Input: Atom coordinates, the atomic number of the
atoms, electron basis set
• Calculate Coulombic and exchange contributions
• Form Fock matrix using first guess orbitals
• Diagonalize Fock matrix and obtain improved
energies and orbitals
• Repeat cycle until energies and orbitals remain
unchanged
• Output: Energy, forces, electronic structure, etc.
Recall that the electron WF is characterized by spatial and spin
variables
• If the number of electrons is even and orbitals are doubly
occupied, we have a closed shell (Fig. a)
• If the number of electrons is odd, we have an open-shell
system (Fig. b)
• In general, if the numbers of electrons with spins up and
down are different, we have an open-shell system
Open-shell systems can be dealt with by one of two HartreeFock methods:
• Restricted open-shell Hartree–Fock (ROHF)
• Unrestricted Hartree–Fock (UHF)
UHF theory is the most commonly used when the number of electrons of each spin are not equal.
While RHF theory uses a single molecular orbital twice, UHF theory uses different molecular orbitals
for the 𝛼 and 𝛽 electrons. The result is a pair of coupled Roothaan equations
𝐅 𝛼 𝐜 𝛼 = 𝐒𝐜 𝛼 𝜖 𝛼
𝐅𝛽 𝐜 𝛽 = 𝐒𝐜 𝛽 𝜖 𝛽
The pair of equations are coupled because each orbital has to be optimised in the average field of all
other electrons. This yields sets of molecular orbitals and orbital energies for both the 𝛼 and 𝛽 spin
electrons.
UHF method has one drawback.
• A single Slater determinant of different orbitals for different spins is not a satisfactory
eigenfunction of the total spin operator -𝐒 2
• The electronic state can be contaminated by excited states. For example, the doublet state (one
more 𝛼 spin electron than 𝛽 spin) may have too large total spin eigenvalue. If 𝐒 2 = 0.8 or less
1 1
(exact value is + 1 = 0.75), it is probably satisfactory. If it is 1.0 or so, it is certainly not
2 2
satisfactory and the calculation should be rejected and a different approach taken.
Given this drawback, why is the UHF used in preference to the ROHF?
• UHF is simpler to code
• it is easier to develop post-Hartree–Fock methods
• it is unique, unlike ROHF where different Fock operators can give the same final wave function.
Water dissociation
• Large systematic error in HF electronic
energy
• The equilibrium structure and the shape of
the potential energy surface near the
equilibrium position is fairly well described
• The RHF method fails when atoms gain ionic
character
• The UHF can be used to calculate relative
energy changes when molecule dissociates
• Helium atom Hamiltonian
∇12 ∇22 2 2 1
𝐻=− − − − +
2 2 𝑟1 𝑟2 𝑟12
• True wavefunction posesses a coulomb hole at 𝑟1 =
𝑟2 and 𝛼 = 0.
• The HF wavefunction (expressed as a single Slater
determinant) has no special behaviour near
coalescence, i.e. No electron correlation
• In reality, electrons tend to avoid each other.
• Unphysically large propabilities to finding two
electrons near eachother results in overestimated
potential energies in Hartree-Fock calculations
Lack of electron correlation in HF leads to an error of 1 eV (100 kJ/mol) per (valence)electron pair.
• In some cases the errors cancel themselves out
• Isodesmic reactions wherein the number and type of bonds on each side of the reaction remains
unchanged can be calculated within few kcal/mol
Example: Reaction of ethanol with methane
• ethanol + methane → methanol + ethane
• The heat of formation of ethanol can be estimated now by simply calculating the reaction energy,
with quantum mechanical methods and by using the computed reaction energy together with the
known heats of formation.
• Using cheaply calculated HF/STO-3G energies for all four species, a reaction energy of +10.9
kJ/mol is predicted. Together with the know heats of formation, a value of -220 kJ/mol is predicted
for ethanol. This has to be compared to the experimental value of -235.3 kJ/mol.
Hartree-Fock and experimental equilibrium bond lengths 𝑅𝑒 (in pm)
Molecule HF Exp.
F2 -155.3 163.4
H2 350.8 458.0
HF 405.7 593.2
H2 O 652.3 975.3
O3 -238.2 616.2
CO2 1033.4 1632.5
C2 H4 1793.9 2359.8
CH4 1374.1 1759.3
Reaction HF Exp.
CO + H2 → CH2 O 2.7 -21.8
H2 O + F2 → HOF + HF -139.1 -129.4
N2 +3H2 → 2NH3 -147.1 -165.4
C2 H2 +H2 → C2 H4 -214.1 -203.9
CO2 +4H2 → CH4 + 2H2 O -242.0 -245.3
2CH2 → C2 H4 -731.8 -845.7
O3 +3H2 → 3H2 O -1142.7 -935.5
nucleus nucleus
electron 2 electron 2
the wave function of electron 1 while keeping the another fixed at x=0.5
Helgaker, Jorgensen, Olsen: Molecular Electronic-Structure Theory (Wiley 2002)
Correlation hole
Difference between the exact and HF wave functions for the He atom
nucleus
electron 2
Configuration interaction method
Configuration interaction (CI) has the following characteristics:
• A post-HartreeFock linear variational method.
• Solves the nonrelativistic Schrödinger equation within the BO–approximation for a multi-
electron system.
• Describes the linear combination of Slater determinants used for the wave function.
o Orbital occupation (for instance, 1s2 2s2 1p1 ...) interaction means the mixing of
different electronic configurations (states).
In contrast to the HartreeFock method, in order to account for electron correlation, CI uses a
variational wave function that is a linear combination of configuration state functions (CSFs)
built from spin orbitals:
𝜓= 𝐶𝑖 𝜓𝑖 = 𝐶0 𝜓0 + 𝐶1 𝜓1 + 𝐶2 𝜓2 + ⋯
𝑖
Coefficients from the wavefunction expansion are determined by a variational optimization respect
to the electronic energy
𝐇𝐂 = 𝐄𝐂𝐈 𝐂
where H is the Hamiltonian matrix with matrix elements
𝐻𝑖𝑗 = 𝜓𝑖 𝐻 𝜓𝑗
The construction of the CI wavefunction may be carried out by diagonalization of the Hamiltonian
matrix, but in reality iterative techniques are used to extract eigenvalues and eigenfunctions
(Newton’s method).
The first term in the CI-expansion is normally the Hartree–Fock determinant
𝜓 = 𝐶0 𝜓HF + 𝐶1 𝜓1 + 𝐶2 𝜓2 + ⋯
The other CSFs can be characterised by the
number of spin orbitals that are swapped with
virtual orbitals from the Hartree–Fock
determinant
• If only one spin orbital differs, we describe this
as a single excitation determinant
• If two spin orbitals differ it is a double
excitation determinant and so on
• The eigenvalues are the energies of the ground
and some electronically excited states. By this
it is possible to calculate energy differences
(excitation energies) with CI-methods.
The expansion to the full set of Slater determinants (SD) or CSFs by distributing all electrons among
all orbitals is called full CI (FCI) expansion. FCI exactly solves the electronic Schrödinger equation
within the space spanned by the one-particle basis set.
• In FCI, the number SDs increase very rapidly with the number of electrons and number of orbitals.
For example, when distributing 10 electrons to 10 orbitals the number of SDs is 63504. This
illustrates the intractability of the FCI for any but the smallest electronic systems.
• Practical solution: Truncation of the CI-expansion. Truncating the CI-space is important to save
computational time. For example, the method CID is limited to double excitations only. The
method CISD is limited to single and double excitations. These methods, CID and CISD, are in
many standard programs.
CI-expansion truncation is handled differently between static or dynamical correlation.
• In the treatment of static correlation in addition to the dominant configurations, near degenerate
configurations are chosen (referred as reference configurations).
• Dynamical correlation is subsequently treated by generating excitations from reference space.
Excitation energies of truncated CI-methods are generally too high because the excited states are
not that well correlated as the ground state is.
• The Davidson correction can be used to estimate a correction to the CISD energy to account for
higher excitations. It allows one to estimate the value of the full configuration interaction energy
from a limited configuration interaction expansion result, although more precisely it estimates
the energy of configuration interaction up to quadruple excitations (CISDTQ) from the energy of
configuration interaction up to double excitations (CISD). It uses the formula:
δ𝐸𝑄 = 1 − 𝐶02 𝐸CISD − 𝐸HF
where 𝐶0 is the coefficient of the Hartree–Fock wavefunction in the CISD-expansion
CI-methods are not size-consistent and size-extensive
• Size-inconsistency means that the energy of two infinitely separated particles is not double the
energy of the single particle. This property is of particular importance to obtain correctly
behaving dissociation curves.
• Size-extensivity, on the other hand, refers to the correct (linear) scaling of a method with the
number of electrons.
• The Davidson correction can be used.
• Quadratic configuration interaction (QCI) is an extension CI that corrects for size-consistency
errors in the all singles and double excitation CI methods (CISD). This method is linked to
coupled cluster (CC) theory.
o Accounts for important four-electron correlation effects by including quadruple excitations
Example: CI calculation for Helium atom
Lets us begin with two-configuration wavefunction expressed as a linear combination of hydrogenic
wavefunctions having the form
𝜓1,2 = 𝑐1 𝜓1 + 𝑐2 𝜓2
where 𝜓1 arises from the configuration 1𝑠 2 and 𝜓2 arises from the configuration 1𝑠2𝑠. Specifically,
the two wavefunctions are
1 1𝑠 1 𝛼 1 1𝑠 1 𝛽 1
𝜓1 =
2 1𝑠 2 𝛼 2 1𝑠 2 𝛽 2
1 1𝑠 1 𝛼 1 1𝑠 1 𝛽 1 1 2𝑠 1 𝛼 1 2𝑠 1 𝛽 1
𝜓2 = +
2 2𝑠 2 𝛼 2 2𝑠 2 𝛽 2 2 1𝑠 2 𝛼 2 1𝑠 2 𝛽 2
Since both 𝜓1 and 𝜓2 describe singlet states, there will be no vanishing matrix elements of 𝐻.
If we represent these matrix elements by 𝐻𝑖𝑗 = 𝜓𝑖 𝐻 𝜓𝑗 (𝑖, 𝑗 = 1 or 2), the secular
determinant to be solved is
𝐻11 − 𝐸 𝐻12
=0
𝐻12 𝐻22 − 𝐸
The diagonal matrix elements 𝐻11 and 𝐻22 are just the energies of single configurational
calculations for the ground and exited states. Keeping in mind that the spin portions integrate
out separately to unity 𝛼 𝛼 = 𝛽 𝛽 = 1 , 𝛼 𝛽 = 0 we obtain for the 𝐻11
𝐻11 = 1𝑠 1 1𝑠 2 𝐻 1𝑠 1 1𝑠 2 = 2𝜖1 + 𝐽11
where we represent the 1𝑠 orbital by subscript 1. The energy 𝜖 has the same form as the
hydrogen atom energy
𝑍2 2
𝜖𝑛 = − 2 = − 2
2𝑛 𝑛
Similarly we can obtain the expressions for the matrix elements 𝐻22 and 𝐻12 but this derivation
is omitted here for the sake of simplicity and only the final results are shown. After the
appropriate integrations, the matrix elements become
𝐻11 = −2.75
𝐻22 = −2.037
𝐻12 = 0.253
The roots of the quadratic formula (produced by secular determinant) are 𝐸1 = −2.831 and
𝐸2 = −1.956. The lower root represents the ground state whose experimental energy is
−2.903. Note that this improves the single configurational result −2.75. The higher root
represents the lowest excited state (experimental energy = −2.15).
Compared to CISD-method, the simpler and less computationally expensive MP2-method gives
superior results when size of the system increases (MP2 is size extensive).
• For water monomer, MP2 recovers 94% of correlation energy which remains similar with increasing
system (cc-pVTZ basis).
• For stretched water monomer (bond length doubled) CISD recovers only 80.2% of the correlation
energy.
o for a large variety of systems it recovers 80-90% of the available correlation energy
o With the Davidson correction added, the error is reduced to 3%.
• When the number of monomers increases, the degradation in the performance is even more
severe for the equilibrium geometry.
o For eight monomers, the CISD wavefunction recovers only half of the correlation energy and
the Davidson correction remain more or less the same.
Dissociation of a water molecule
Thick line: Full CI
One the right: Difference between
truncated CI and FCI
In the picture (a) are shown the energy levels of harmonic oscillator. Its eigenfunctions and energy
levels are known. In the picture (b) the system is perturbed by known potential energy term V’ which
is the difference between potential energy curves of systems (a) and (b). The eigenfunctions and
energy levels of system (b) are not known. The perturbation theory is devised to found approximate
solutions for the properties of the perturbed system.
The Schrödinger equation for the perturbed state n is
𝐻𝜓𝑛 = 𝐻 0 + 𝜆𝑉 𝜓𝑛 = 𝐸𝑛 𝜓𝑛
0 𝑖 𝑖 𝑖
𝐻 + 𝜆𝑉 𝜆𝑖 𝜓𝑛 = 𝜆𝑖 𝐸𝑛 𝜆𝑖 𝜓𝑛
𝑖=0 𝑖=0 𝑖=0
Writing only the first terms we obtain
0 0 1 0 1 0 1
𝐻 + 𝜆𝑉 𝜓𝑛 + 𝜆𝜓𝑛 = 𝐸𝑛 + 𝜆𝐸𝑛 𝜓𝑛 + 𝜆𝜓𝑛
The zeroth-order equation is simply the Schrödinger equation for the unperturbed system
0 0 0 0
𝐻 𝜓𝑛 = 𝐸𝑛 𝜓𝑛
The first-order terms are those which are multiplied by 𝜆
0 1 0 0 1 1 0
𝐻 𝜓𝑛 + 𝑉𝜓𝑛 = 𝐸𝑛 𝜓𝑛 + 𝐸𝑛 𝜓𝑛
0 ∗
When this is multiplied through by 𝜓𝑛 from left and integrated, the first term on the left-hand
side cancels with the first term on the right-hand side (The 𝐻 0 is hermitian). This leads to the first-
order energy shift:
1 0 0
𝐸𝑛 = 𝜓𝑛 𝑉 𝜓𝑛
This is simply the expectation value of the perturbation Hamiltonian while the system is in the
0 1
unperturbed state. The energy of the 𝑛th state up to the first order is thus 𝐸𝑛 + 𝐸𝑛 .
Interpretation of the first order correction to energy:
1 0 0
𝐸𝑛 = 𝜓𝑛 𝑉 𝜓𝑛
0
• The perturbation is applied, but we keep the system in the quantum state 𝜓𝑛 , which is a valid
quantum state though no longer an energy eigenstate.
0 0
• The perturbation causes the average energy of this state to increase by 𝜓𝑛 𝑉 𝜓𝑛
Further shifts are given by the second and higher order corrections to the energy.
To obtain the first-order correction to the energy eigenstate, we recall the expression derived earlier
0 1 0 0 1 1 0
𝐻 𝜓𝑛 + 𝑉𝜓𝑛 = 𝐸𝑛 𝜓𝑛 + 𝐸𝑛 𝜓𝑛
0 ∗
and multiply it by 𝜓𝑚 , 𝑚 ≠ 𝑛 from left and integrate. We obtain
0 0 0 1 0 0
𝐸𝑚 − 𝐸𝑛 𝜓𝑚 𝜓𝑛 = − 𝜓𝑚 𝑉 𝜓𝑛
0
Multiplying from left with 𝑚≠𝑛 𝜓𝑚 and using the resolution of identity, we obtain
0 0
𝜓𝑚 𝑉 𝜓𝑛
1 0
𝜓𝑛 = 0 0
𝜓𝑚
𝑚≠𝑛 𝐸𝑛 − 𝐸𝑚
The first-order change in the 𝑛-th energy eigenfunction has a contribution from each of the energy
eigenstates 𝑚 ≠ 𝑛.
The second order correction to the energy is
2
0 0
𝜓𝑚 𝑉 𝜓𝑛
2
𝐸𝑛 = 0 0
𝑚≠𝑛 𝐸𝑛 − 𝐸𝑚
Conclusion:
0 0
• Each term is proportional to the matrix element 𝜓𝑚 𝑉 𝜓𝑛 , that is a measure of how much the
0 0
perturbation mixes state 𝜓𝑛 with state 𝜓𝑚
0 0
• Correction is inversely proportional to the energy difference between states 𝜓𝑛 and 𝜓𝑚 , which
means that the perturbation deforms the state to a greater extent if there are more states at nearby
energies.
0
• Expression is singular if any of these states have the same energy as state 𝜓𝑛 , which is why we
assumed that there is no degeneracy
• Higher-order deviations can be found by a similar procedure
Example: PT calculation for Helium atom
Helium atom Hamiltonian
∇12 ∇22 2 2 1 1
𝐻=− − − − + = ℎ1 + ℎ2 +
2 2 𝑟1 𝑟2 𝑟12 𝑟12
It is convinient to choose the unperturbed system a a two-electron atom in which the electrons do not
interact. The zeroth_order Hamiltonian then is
𝐻 0 = ℎ1 + ℎ2
The zeroth-order eigenfunctions have the hydrogen atom form
3 2
0
1 𝑍
𝜓 = 1𝑠 1 1𝑠 2 = 𝑒 −𝑍 𝑟1 +𝑟2 𝑎0
𝜋 𝑎0
And the zeroth-order energy is simply
𝐸 0 = −𝑍 2 = −4.0
The first-order perturbation-energy correction is
1 1 5 5
𝐸 = 1𝑠 1 1𝑠 2 1𝑠 1 1𝑠 2 = 𝑍=
𝑟12 8 4
Thus, the total energy of the helium atom (to the first order) is
0 1
5
𝐸=𝐸 +𝐸 = −4 + = −2.75
4
This result is equal to the single configurational energy in previous CI calculation. Thus, electron
correlation does not seem to contribute to the first order energy correction.
The total energy up to the second order is
𝐸=𝐸 0 +𝐸 1 +𝐸 2 = −2.75 − 0.157666405 = −2.90767
which is lower than true ground state energy −2.90372.
Møller–Plesset perturbation theory
Introduction:
The Møller–Plesset perturbation theory (MP) was published as early as 1934 by Christian Møller
and Milton S. Plesset.
• The starting point is eigenfunction of the Fock-operator.
• It improves on the Hartree–Fock method by adding electron correlation effects.
• MP theory is not variational. Calculated energy may be lower than true ground state energy.
MP methods (MP2, MP3, MP4, ...) are implemented in many computational chemistry codes.
Higher level MP calculations, generally only MP5, are possible in some codes. However, they are
rarely used because of their costs.
The MP-energy corrections are obtained with the perturbation 𝑉, which is defined as a difference
between the true nonrelativistic Hamiltonian 𝐻 and the sum of one-electron Fock operators 𝐹
𝑉 =𝐻−𝐹
The Slater determinant 𝜓 is the eigenfunction of the Fock-operator 𝐹
𝐹𝜓 = 𝜖𝑖 𝜓
𝑖
where 𝜖𝑖 is the orbital energy belonging to the doubly occupied space orbital.
The sum of MP zeroth-order energy and first order energy correction is
(0) (1)
𝐸𝑀𝑃 + 𝐸𝑀𝑃 = 𝜓 𝐹 𝜓 + 𝜓 𝑉 𝜓 = 𝜓 𝐻 𝜓
But 𝜓 𝐻 𝜓 is the variational integral for the Hartree-Fock wave function 𝜓 and it therefore equals
the Hartree-Fock energy 𝐸𝐻𝐹 .
(0) (1)
𝐸𝑀𝑃 + 𝐸𝑀𝑃 = 𝐸𝐻𝐹
In order to obtain the MP2 formula for a closed-shell molecule, the second-order correction formula is written on basis of
doubly-excited Slater determinants (singly-excited Slater determinants vanish).
−1 −1 2
(2) 𝜙𝑎 1 𝜙𝑏 2 𝑟12 𝜙𝑖 1 𝜙𝑗 2 − 𝜙𝑎 1 𝜙𝑏 2 𝑟12 𝜙𝑗 1 𝜙𝑖 2
𝐸𝑀𝑃 =
𝜖𝑖 + 𝜖𝑗 − 𝜖𝑎 − 𝜖𝑏
𝑖>𝑗
𝑎>𝑏
where 𝜙𝑖 and 𝜙𝑗 are occupied orbitals and 𝜙𝑎 and 𝜙𝑏 are virtual (unoccupied) orbitals. The quantities 𝜖𝑖 , 𝜖𝑗 , 𝜖𝑎 , and 𝜖𝑏
are the corresponding orbital energies. Up to the second-order, the total electronic energy is given by the Hartree–Fock
energy plus second-order MP correction:
(2)
𝐸 = 𝐸𝐻𝐹 + 𝐸𝑀𝑃
Calculated and experimental atomization energies (kJ/mol)
𝜓 = 𝑒 𝑇 𝜙0
Where is a Slater determinant usually constructed from Hartree–Fock molecular orbitals. The
operator 𝑇 is an excitation operator which, when acting on 𝜙0 , produces a linear combination of
excited Slater determinants.
The exponential approach guarantees the size extensivity of the solution. For two subsystems A and
B and corresponding excitation operators 𝑇A and 𝑇B , the exponential function admits for the simple
factorization 𝑒 𝑇A +𝑇B = 𝑒 𝑇A 𝑒 𝑇B . Therefore, aside from other advantages, the CC method maintains
the property of size consistency.
The cluster operator is written in the form
𝑇 = 𝑇1 + 𝑇2 + 𝑇3 + ⋯
where 𝑇1 is the operator of all single excitations, 𝑇2 is the operator of all double excitations and so
forth.
The exponential operator 𝑒 𝑇 may be expanded into Taylor series:
2 2 2
𝑇 𝑇1 𝑇2
𝑒𝑇 = 1 + 𝑇 + + ⋯ = 1 + 𝑇1 + 𝑇2 + + 𝑇1 𝑇2 + +⋯
2! 2 2
In practice the expansion of 𝑇 into individual excitation operators is terminated at the second or
slightly higher level of excitation. Slater determinants excited more than 𝑛 times contribute to the
wave function because of the non-linear nature of the exponential function. Therefore, coupled
cluster terminated at 𝑇𝑛 usually recovers more correlation energy than CI with maximum 𝑛
excitations.
A drawback of the method is that it is not variational
𝜙 𝑒 −𝑇 𝐻𝑒 𝑇 𝜙
𝐸𝜙 =
𝜙𝜙
which for truncated cluster expansion becomes
𝜒𝐻𝜊
𝐸Φ =
𝜙𝜙
where 𝜒 and 𝜊 are different functions
The classification of traditional coupled-cluster methods rests on the highest number of excitations
allowed in the definition of 𝑇. The abbreviations for coupled-cluster methods usually begin with
the letters CC (for coupled cluster) followed by
• S - for single excitations (shortened to singles in coupled-cluster terminology)
• D - for double excitations (doubles)
• T - for triple excitations (triples)
• Q - for quadruple excitations (quadruples)
Thus, the operator in CCSDT has the form:
𝑇 = 𝑇1 + 𝑇2 + 𝑇3
Terms in round brackets indicate that these terms are calculated based on perturbation theory. For
example, a CCSD(T) approach simply means:
• It includes singles and doubles fully
• Triples are calculated with perturbation theory.
Calculated and experimental atomization energies (kJ/mol)
Method 𝒓𝐞 1.5 𝒓𝐞 2 𝒓𝐞
Hartree Fock 216.1 270.9 370.0
CID 13.7 34.5 84.8
CISD 12.9 30.4 75.6
CISDT 10.6 23.5 60.3
CISDTQ 0.40 1.55 6.29
CCD 5.01 15.9 40.2
CCSD 4.12 10.2 21.4
CCSDT 0.53 1.78 -2.47
CCSDTQ 0.02 0.14 -0.02
• Usually these functions are centered on atoms, but functions centered in bonds or lone pairs have been used.
• Additionally, basis sets composed of sets of plane waves are often used, especially in calculations involving
systems with periodic boundary conditions (continuous systems, surfaces).
Quantum chemical calculations are typically performed within a finite set of basis functions.
• These basis functions are usually not the exact atomic orbitals, like the hydrogen atom eigenfunctions.
• If the finite basis is expanded towards an infinite complete set of functions, calculations using such a basis set
are said to approach the basis set limit.
In the early days of quantum chemistry so-called Slater type orbitals (STOs) were used as basis functions
due to their similarity with the eigenfunctions of the hydrogen atom. Their general definition is
𝜓𝑛𝑙𝑚 𝑟, 𝜃, 𝜙 = 𝑁𝑟 𝑛−1 𝑒 −𝜁𝑟 𝑎0 𝑌 𝑚 𝜃, 𝜙
𝑙
where 𝑛 = 1,2, … is related to hydrogen atom principal quantum number, and 𝑙 and 𝑚 are related to
hydrogen atom angular momentum and magnetic quantum numbers, respectively. 𝑁 is a normalization
factor, 𝜁 is the effective nuclear charge, and 𝑌𝑙𝑚 𝜃, 𝜙 being the spherical harmonics.
• STOs have an advantage in that they have direct physical interpretation and thus are naturally good
basis for molecular orbitals.
• From a computational point of view the STOs have the severe shortcoming that most of the required
integrals needed in the course of the SCF procedure must be calculated numerically which drastically
decreases the speed of a computation.
• Still, today there exist some modern and efficient computational chemistry program packages that use
STOs (ADF).
STOs can be approximated as linear combinations of Gaussian type orbitals, which are defined as
𝑖 𝑗 𝐫−𝐑 2
𝑔𝑖𝑗𝑘 𝐫 = 𝑁 𝑥 − 𝑅𝑥 𝑦 − 𝑅𝑦 𝑧 − 𝑅𝑧 𝑘 𝑒 −𝛼
𝑁 is a normalization factor, 𝐑 is the atomic center, and 𝛼 is an orbital exponent of the Gaussian
function, respectively. GTOs are not really orbitals, they are simpler functions (Gaussian primitives).
GTOs are usually obtained from quantum calculations on atoms (i.e. Hartree-Fock or Hartree-Fock
plus some correlated calculations, e.g. CI).
• Typically, the exponents 𝛼 are varied until the lowest total energy of the atom is achieved.
• For molecular calculations, certain linear combinations of GTOs will be used as basis functions.
• Such a basis function (contraction) will have its coefficients and exponents fixed. For example:
𝜙1 = 𝑎𝑔1 + 𝑏𝑔2 + 𝑐𝑔3
Where coefficients 𝑎, 𝑏, and 𝑐 and the exponents 𝛼 in functions 𝑔 are fixed (i.e. are not variables).
2
The main difference to the STOs is that the variable 𝑟 in the exponential function 𝑒 −𝛼 𝐫−𝐑 is
squared. Generally the inaccuracy at the center or the qualitatively different behaviour at
long distances from the center have a marked influence on the results.
where
𝛼a 𝐑 a + 𝛼c 𝐑 c
𝐑′ =
𝛼a + 𝛼c
and
𝛼a 𝛼c
𝛼′ =
𝛼a + 𝛼c
Thus, the product of two GTOs on different centers is equal to a single other GTO at center R′ between
the original centers. As a result, even a four-center integral over GTOs can be written as two-center two-
electron integral. A similar reduction does not arise for STOs.
𝑗 2
In GTOs 𝑁 𝑥 − 𝑅𝑥 𝑖 𝑦 − 𝑅𝑦 𝑧 − 𝑅𝑧 𝑘 𝑒 −𝛼 𝐫−𝐑 the sum of the exponents of the cartesian
coordinates, 𝐿 = 𝑖 + 𝑗 + 𝑘, is used to mark functions as 𝑠-type (𝐿 = 0), 𝑝-type (𝐿 = 1), 𝑑-type (𝐿 = 2),
and so on
• Unfortunately GTOs are not eigenfunctions of the squared angular momentum operator 𝐿2 .
• However, combinations of GTOs are able to approximate correct nodal properties of atomic orbitals
by taking them with different signs. For example combining three 𝑑-type cartesian GTOs yields a
cartesian GTO of 𝑠-type:
𝑔200 + 𝑔020 + 𝑔002 ∝ 𝑔000
• Today, there are hundreds of basis sets composed of GTOs. The smallest of these are called minimal
basis sets, and they are typically composed of the minimum number of basis functions required to
represent all of the electrons on each atom. The largest of these can contain literally dozens to
hundreds of basis functions on each atom.
A minimum basis set is one in which a single basis function is used for each orbital in a Hartree-Fock
calculation on the free atom.
• The most common minimal basis set is STO-𝑛G, where 𝑛 is an integer. This n value represents the
number GTOs used to approximate STO for both core and valence orbitals.
• Minimal basis sets typically give rough results that are insufficient for research-quality publication,
but are much cheaper than their larger counterparts.
• Commonly used minimal basis sets of this type are: STO-3G, STO-4G, STO-6G
• Example: For lithium, GTOs of 𝑝-type are added to the basis functions corresponding to the 1𝑠
and 2𝑠 orbitals of the free atom.
The minimal basis sets are not flexible enough for
accurate representation of orbitals
Solution: Use multiple functions to represent each
orbital
For example, the double-zeta basis set allows us to
treat each orbital separately when we conduct the
Hartree-Fock calculation.
𝜓2𝑠 𝐫 = 𝑐1 𝜓2𝑠 𝐫; 𝜁1 + 𝑐2 𝜓2𝑠 𝐫; 𝜁2
where 2𝑠 atomic orbital is expressed as the sum of
two STOs. The 𝜁-coefficients account for how large
the orbitals are. The constants 𝑐1 and 𝑐2 determine
how much each STO will count towards the final
orbital.
The triple and quadruple-zeta basis sets work the same way, except use three and four Slater
equations (linear combination of GTOs) instead of two. The typical trade-off applies here as
well, better accuracy...more time/work.
There are several different types of extended basis sets
• split-valence
• polarized sets
• diffuse sets
• correlation consistent sets
Pople’s split-valence basis sets n-ijG or n-ijkG.
• n - number of GTOs for the inner shell orbitals; ij or ijk
– number of GTOs for basis functions in the valence
shell. The ij notations describes sets of valence double
zeta quality and ijk sets of valence triple zeta quality.
• The 𝑠-type and 𝑝-type functions belonging to the same
electron shell are folded into a 𝑠𝑝-shell. In this case,
number of 𝑠-type and 𝑝-type GTOs is the same, and
they have identical exponents. However, the
coefficients for 𝑠-type and 𝑝-type basis functions are
different.
Example: Four 𝑠-type GTOs used to represent 1𝑠 orbital of hydrogen as:
2 2 2 2
𝜓1𝑠 = 0.50907𝑁1 𝑒 −0.123317𝑟 + 0.47449𝑁2𝑒 −0.123317𝑟 + 0.13424𝑁3 𝑒 −0.123317𝑟 + 0.01906𝑁4 𝑒 −0.123317𝑟
where 𝑁𝑖 is a normalization constant for a given GTO. These GTOs may be grouped in 2 basis functions. The first
basis function contains only 1 GTO:
2
𝜙1 = 𝑁1 𝑒 −0.123317𝑟
3 GTOs are present in the second basis function:
2 2 2
𝜙2 = 𝑁 0.47449𝑁2𝑒 −0.123317𝑟 + 0.13424𝑁3 𝑒 −0.123317𝑟 + 0.01906𝑁4 𝑒 −0.123317𝑟
where 𝑁 is a normalization constant for the whole basis function. In this case, 4 GTOs were contracted to 2 basis
functions. It is frequently denoted as 4𝑠 → 2𝑠 contraction. The coefficients in function are then fixed in
subsequent molecular calculations.
Example: Silicon 6-31G basis set
• The corresponding exponents for 𝑠-type and 𝑝-
type basis functions are equal but coefficients in
𝑠-type and 𝑝-type basis functions are different.
• GTOs are normalized here since coefficients for
basis functions consisting of one GTO (last row)
are exactly 1.
• The basis set above represents the following
contraction 16𝑠, 10𝑝 → 4𝑠, 3𝑝
Example: 3-21G basis set of carbon
Surface and contour plot of p–type basis function including two Gaussians.
𝛼sp,1 : = 3.664980
𝜙2𝑝𝑦 : = 𝑦 ∗ 𝑐2𝑝,1 ∗ 𝜙2𝑝,1 + 𝑐2𝑝,2 ∗ 𝜙2𝑝,2 𝛼sp,2 ≔ 0.770545
3
2∗ 𝛼sp,1 4 𝑥2 +𝑦 2 +𝑧 2
𝜙2𝑝,1 : = ∗ e−𝛼sp,1 𝑐2𝑝,1 : = 0.236460
𝜋 𝑐2𝑝,2 : = 0.860619
Polarized basis sets
• Polarization functions denoted in Pople’s sets by an asterisk
• Two asterisks, indicate that polarization functions are also added to light atoms (hydrogen and
helium).
• Polarization functions have one additional node.
• For example, the only basis function located on a hydrogen atom in a minimal basis set would
be a function approximating the 1𝑠 atomic orbital. When polarization is added to this basis set,
a 𝑝-type function is also added to the basis set.
6-31G**
Polarization functions add flexibility within the basis set, effectively allowing molecular orbitals to be
more asymmetric about the nucleus.
• This is an important for accurate description of bonding between atoms, because the precence of
the other atom distorts the environment of the electrons and removes its spherical symmetry.
• Similarly, 𝑑-type functions can be added to a basis set with valence 𝑝-type orbitals, and so on.
• High angular momentum polarization functions (𝑑, 𝑓, …) are important for heavy atoms
Some observations concerning polarization functions:
• The exponents for polarization functions cannot be derived from Hartree-Fock calculations for
the atom, since they are not populated.
• In practice, these exponents are estimated ”using well established rules of thumb” or by using a
test set of molecules.
• The polarization functions are important for reproducing chemical bonding.
• They should be included in all calculations where electron correlation is important.
• Adding them is costly. Augmenting basis set with 𝑑-type polarization functions adds 5 basis
function on each atom while adding 𝑓-type functions adds 7.
The basis sets are also frequently augmented with the so-called diffuse functions.
• These Gaussian functions have very small exponents and decay slowly with distance from the
nucleus.
• Diffuse gaussians are usually of 𝑠-type and 𝑝-type.
• Diffuse functions are necessary for correct description of anions and weak bonds (e.g. hydrogen
bonds) and are frequently used for calculations of properties (e.g. dipole moments, polarizabilities,
etc.).
For the Pople’s basis sets the following notaton is used:
• n-ij+G, or n-ijk+G when 1 diffuse 𝑠-type and 𝑝-type gaussian with the same exponents are added to
a standard basis set on heavy atoms.
• The n-ij++G, or n-ijk++G are obtained by adding 1 diffuse 𝑠-type and 𝑝-type gaussian on heavy atoms
and 1 diffuse 𝑠-type gaussian on hydrogens.
Diffuse functions
Example: For a helium atom we can choose the trial function as follows:
𝜙 𝑟1 , 𝑟2 , 𝑟12 = 𝐶 1 + 𝑝𝑟12 + 𝑞 𝑟1 − 𝑟2 2 exp −𝛼 𝑟1 + 𝑟2
where 𝐶 is normalization constant and 𝑝, 𝑞, and 𝛼 are variational parameters.
After optimization:
𝑝 = 0.30, 𝑞 = 0.13, and 𝛼 = 1.816
𝐸 = −2.9024 (Three parameters)
𝐸 = −2.9037462 (1024 parameters)
𝐸 = −2.9037843 (Experimental value)
Explicitly correlated methods
• R12 and F12 methods include two-electron basis functions that depend on the interparticle
distance 𝑟12
• These theories bypass the slow convergence of conventional methods
• High accuracy can be achieved dramatically faster
• With these methods it has been achieved kJ/mol accuracy for molecular systems consisting of up
to 18 atoms
• This result is well below the so called chemical accuracy, that is, an error of 1 kcal/mol (4.184
kJ/mol)
• Methods: CCSD-R12 or F12, CCSD(T)-R12 or F12, MP2-R12 or F12
Example: Calculated and experimental geometric parameters
𝜌1 𝐫 = 𝑁 … ψ 𝐫1 , 𝐫2 , … , 𝐫𝑁 2 𝑑𝜎 𝑑𝐫 … 𝑑𝐫
1 2 𝑁
𝜌 𝐫 = 𝜌1 𝐫 𝑑𝜎1
o 𝜌1 describes the probability of finding any of the 𝑁 electrons within the volume element in
the spin state 𝜎, with the other 𝑁-1 electrons having arbitrary positions and spin states
o 𝜌 is an observable (e.g. X-ray spectroscopy)
• Properties
𝜌 𝐫→∞ =0
𝜌 𝐫 𝑑𝐫1 = 𝑁
Pair density
• Let’s generalize: the probability for finding two electrons in spin states and in the volume
elements and is given by the pair density
𝜌2 𝐫1 , 𝐫2 = 𝑁 𝑁 − 1 … ψ 𝐫1 , 𝐫2 , … , 𝐫𝑁 2 𝑑𝜎 𝑑𝜎 𝑑𝐫 … 𝑑𝐫
1 2 3 𝑁
𝑃 𝐫1 , 𝐫2 = 𝜌2 𝐫1 , 𝐫2 𝑑𝜎1 𝑑𝜎2
• The pair density contains all information about electron correlation, and we can express the
energy of any system in any state as
1 𝜌 𝐫1 1 𝑃 𝐫1 , 𝐫2
𝐸=− ∇2 𝜌 𝐫1 𝑑𝐫1 + 𝑑𝐫1 + 𝑑𝐫1 𝑑𝐫2
2 𝐫1 − 𝐑 𝑖 2 𝐫1 − 𝐫2
𝑖
• Rewrite
𝜌2 𝐫1 , 𝐫2 = 𝜌1 𝐫1 𝜌1 𝐫2 1 + 𝑓 𝐫1 , 𝐫2
Where 𝑓 𝐫1 , 𝐫2 is called correlation factor. For example 𝑓 = 0 corresponds to uncorrelated case.
• The difference between the probability to find any electron in 𝑑𝐫2 while there is an electron in
𝑑𝐫1 and uncorrelated case is proportional to the exchange correlation hole ℎxc 𝐫1 ; 𝐫2
𝜌1 𝐫1 ℎxc 𝐫1 ; 𝐫2 = 𝜌2 𝐫1 , 𝐫2 − 𝜌1 𝐫1 𝜌1 𝐫2 = 𝜌1 𝐫2 𝑓 𝐫1 , 𝐫2
• We observe that
1 𝑃 𝐫1 , 𝐫2 1 𝜌 𝐫1 𝜌 𝐫2 1 𝜌 𝐫1 𝜌1 𝐫2 ℎxc 𝐫1 ; 𝐫2
𝐸ee = 𝑑𝐫1 𝑑𝐫2 = 𝑑𝐫1 𝑑𝐫2 + 𝑑𝐫1 𝑑𝐫2
2 𝐫1 − 𝐫2 2 𝐫1 − 𝐫2 2 𝐫1 − 𝐫2
= 𝐽 𝜌 + 𝐸xc 𝜌
• The ℎxc can be formally splitted into
𝜎 =𝜎2
ℎxc = ℎx 1 𝐫1 ; 𝐫2 + ℎc 𝐫1 ; 𝐫2
where ℎx is exchange hole, due to Pauli principle (wave function antisymmetry) and ℎc is correlation
hole, due to electrostatic repulsion.
• Only the ℎxc can be given proper meaning
• The Hartree-Fock theory accounts for ℎx but neglects ℎc .
The Hohenberg-Kohn theorems
The Hohenberg-Kohn theorems relate to any system consisting of electrons moving under the
influence of an external potential
Theorem 1. The external potential and hence the total energy, is a unique functional of the electron
density
• The first H-K theorem demonstrates that the ground state properties of a many-electron system are
uniquely determined by an electron density that depends on only 3 spatial coordinates.
• It lays the groundwork for reducing the many-body problem of N electrons with 3N spatial
coordinates to 3 spatial coordinates, through the use of functionals of the electron density.
• This theorem can be extended to the time-dependent domain to develop time-dependent density
functional theory (TDDFT), which can be used to describe excited states.
Theorem 2. The groundstate energy can be obtained variationally: the density that minimises the total
energy is the exact ground state density
• The second Hohenberg-Kohn theorem has two drawbacks. Firstly, it assumes that there is no
degeneracy in the ground state, and secondly the density has unknown form.
The uniform electron gas
There is no systematic way to find or improve a density functional. The most appealing way forward is to
find the exact solution for a model system, and then assume that the system of interest behaves similarly
to the model.
• The first density functionals were due to Thomas, Fermi, and Dirac, all of which used the uniform
electron gas as their model.
• The uniform electron gas is defined as a large number of electrons N in a cube of volume V,
throughout which there is a uniform spread of positive charge sufficient to make the system neutral.
The uniform gas is then defined as the limit N → ∞, V → ∞, with the density ρ = N/V remaining finite.
• Although it does bear some resemblance to electrons in metals, its widespread use is due to its
simplicity: It is completely defined by one variable, the electron density ρ. Using the uniform electron
gas, an expression for the kinetic energy (the Thomas-Fermi kinetic functional) can be derived
3
𝑇 TF27 𝜌𝜎 = 6𝜋 2 2 3 𝜌𝜎5 3 𝐫 𝑑𝐫, where σ can take the values of α or β.
10
The importance of simple Thomas-Fermi model is not how well it performs in computing the
ground state energy and density but more as an illustration that the energy can be determined
purely using the electron density.
When applied to atoms and molecules the Thomas-Fermi functional yields kinetic energies that are
about 10% too small. Similarly, an expression for the exchange energy of the uniform electron gas
can be calculated (the Dirac exchange functional)
1 3
3 3
𝐸xD30 𝜌𝜎 =− 𝜌𝜎4 3
𝐫 𝑑𝐫
2 4𝜋
The Dirac functional also gives exchange energies that are roughly 10% smaller than those from HF
theory.
The non-uniform electron gas
The electron densities of atoms and molecules are often far from uniform, so functionals based on
systems which include an inhomogeneous density should perform better. In 1935 von Weizsacker
placed infinitesimally small ripples on the uniform electron gas and calculated the second order
correction to the kinetic energy
1
𝑇 W35 𝜌𝜎 = 𝑇 TF27 𝜌𝜎 + 𝜌𝜎5 3
𝐫 𝑥𝜎2 𝐫 𝑑𝐫
8
The corrected functional gives exchange energies that are typically within 3% of HF; however, it is not
seen as an improvement over the Dirac functional, as the potential is unbounded in the Rydberg
regions of atoms and molecules.
Kohn-Sham DFT
The kinetic energy has a large contribution to the total energy. Therefore even the 1% error in the
kinetic energy of the Thomas-Fermi-Weizsacker model prevented DFT from being used as a
quantitative predictive tool. Thus DFT was largely ignored until 1965 when Kohn and Sham
introduced a method which treated the majority of the kinetic energy exactly.
Key idea: The intractable many-body problem of interacting electrons in a static external potential is
reduced to a tractable problem of non-interacting electrons moving in an effective potential. The
theory is based on the reference system: N noninteracting electrons moving in effective potential
𝑣eff , each in one of N orbitals, 𝜓𝑖 .
The central equation in Kohn-Sham DFT is the one-electron Schrödingerlike equation expressed as:
1
− ∇2𝑖 +𝑣eff 𝐫𝑖 𝜓 𝑖 = 𝜖𝑖 𝜓 𝑖
2
The kinetic energy and electron density are given by
1
𝑇s 𝜌 = 𝜓𝑖 − ∇2𝑖 𝜓𝑖
2
𝑖
𝜌 𝐫 = 𝜓𝑖 𝐫 2
LDA 𝜌 =
𝐸xc 𝜖xc 𝜌 𝜌 𝐫 𝑑𝐫
The correlation energy is more complicated and numerous different approximations exist for 𝐸c .
• Strictly, the LDA is valid only for slowly varying densities.
• LDA works surprisingly well with calculations of atoms, molecules, and solids (especially for metals).
o Systematic error cancelation: Typically, in inhomogeneous systems LDA underestimates correlation but
overestimates exchange, resulting in unexpectedly good energy value.
• LDA tends to overestimate cohesive energies by ∼15-20% and underestimates lattice constants by ∼2-3%
for metals and insulators.
• Problem with LDA becomes more severe for weakly bonded systems, such as vdW and H-bonded systems.
o For example, the binding energy of the water dimer and the cohesive energy of bulk ice are both >50%
too large with LDA compared to the experimental values.
o Long range vdW interactions are completely missing in LDA.
Generalized gradient approximation
LDA treats all systems as homogeneous. However, real systems are inhomogeneous. Generalized gradient
approximation (GGA) takes this into account by including the derivative information of the density into the
exchange-correlation functionals.
GGA 𝜌 =
𝐸xc 𝑓 𝜌 𝐫 , ∇𝜌 𝐫 𝑑𝐫
• It is not the physics per se but obtained results that guide the mathematical constructs
o Some successful functionals are not based on any physical considerations
o For example let’s look two popular functionals: In PBE, the functional parameters are obtained from
physical constraints (non-empirical). In B88, functional parameters are obtained from empirical fitting
(empirical).
• GGAs are often called “semi-local” functionals due to their dependence on ∇𝜌 𝐫 .
• In comparison with LDA, GGA tend to improve total energies, atomization energies, energy
barriers and structural energy differences. Especially for covalent bonds and weakly bonded
systems many GGAs are far superior to LDA
o Overall though because of flexibility of a choice of 𝑓 𝜌 𝐫 , ∇𝜌 𝐫 a zoo of GGA functionals
have been developed and depending on the system under study a wide variety of results
can be obtained.
• GGA expand and soften bonds, an effect that sometimes corrects and sometimes overcorrects
the LDA prediction
• Whereas the 𝜖xc 𝜌 (in LDA) is well established, the best choice for 𝑓 𝜌 𝐫 , ∇𝜌 𝐫 is still a
matter of debate
The hybrid functionals
Q: Why bother with making GGA exchange functionals at all – we know that the HF exchange is exact; i.e.
These fourth generation functionals add “exact exchange”calculated from the HF functional to some conventional
treatment of DFT exchange and correlation.
• LDA and GGA exchange and correlation functionals are mixed with a fraction of HF exchange
• The most widely used, particularly in the quantum chemistry community, is the B3LYP functional which employs
three parameters, determined through fitting to experiment, to control the mixing of the HF exchange and
density functional exchange and correlation.
Equilibrium C-C and C=C bond distances (Å)
In practice, DFT can be applied in several distinct ways depending on what is being investigated.
• In solid state calculations, the local density approximations are still commonly used along with plane wave
basis sets, as an electron gas approach is more appropriate for electrons delocalised through an infinite solid.
• In molecular calculations more sophisticated functionals are needed, and a huge variety of exchange-
correlation functionals have been developed for chemical applications.
• In the chemistry community, one popular functional is known as BLYP (from the name Becke for the exchange
part and Lee, Yang and Parr for the correlation part).
• Even more widely used is B3LYP which is a hybrid functional in which the exchange energy, in this case from
Becke’s exchange functional, is combined with the exact energy from Hartree–Fock theory. The adjustable
parameters in hybrid functionals are generally fitted to a ’training set’ of molecules.
• Unfortunately, although the results obtained with these functionals are usually sufficiently accurate for most
applications, there is no systematic way of improving them (in contrast to methods like configuration
interaction or coupled cluster theory)
• Hence in the current DFT approach it is not possible to estimate the error of the calculations without
comparing them to other methods or experiments.
Graphical models
In addition to numerical quantities (bond lengths and angles, energies, dipole moments,…) some
chemically useful information is best displayed in the form of images. For example molecular
orbitals, the electron density, electrostatic potential, etc. These objects can be displayed on screen
using isosurface
𝑓 𝑥, 𝑦, 𝑧 = constant
The constant may be some physical observable of interest, for example, the ”size” of the molecule.
Molecular Orbitals
Chemists are familiar with the molecular orbitals of simple molecules. They recognize
the σ and π orbitals of acetylene, and readily associate these with the molecule’s σ
and π bonds
A simple example where the shape of the highest occupied molecular orbital (HOMO) “foretells” of
chemistry is found in cyanide anion.
- + - +
+ - + -
+ - +
- + -
HOMO in cis-1,3-butadiene is able to interact Interaction between the HOMO on one ethylene
favorably with the LUMO in ethylene (constructive and the LUMO on another ethylene is not favorable, and
overlap) to form cyclohexene concerted addition to form cyclobutane would not be expected
Electron density
Isodensity surfaces may either serve to locate atoms, delineate chemical bonds, or to indicate
overall molecular size and shape.
• The regions of highest electron density surround the heavy (non-hydrogen) atoms in a
molecule. Thus, the X-ray diffraction experiment locates atoms by identifying regions of high
electron density.
• Also interesting, are regions of slightly lower electron density. For example, isodensity surface
(0.1 electrons/au3 ) for cyclohexanone conveys essentially the same information as a
conventional skeletal structure model, that is, it depicts the locations of bonds
A low density surface (0.002 electrons/au3 ), serves to portray overall molecular size and shape. This
is, of course, the same information portrayed by a conventional space-filling (CPK) model.
Bond surfaces (intermediate density) may be applied to elucidate bonding and not only to portray
“known” bonding. For example, the bond surface for diborane clearly shows a molecule with very
little electron density concentrated between the two borons.
This suggests that the appropriate Lewis structure is the one which lacks a boron-boron bond,
rather than the one which has the two borons directly bonded.
Another important application of bond surfaces is to the description of the bonding in transition
states. An example is the pyrolysis of ethyl formate, leading to formic acid and ethylene.
The bond surface offers clear evidence of a “late transition state”. The CO bond is nearly fully
cleaved and the migrating hydrogen is more tightly bound to oxygen (as in the product) than to
carbon (as in the reactant).
Spin density
For open-shell molecules, the spin density indicates the distribution of unpaired electrons. Spin
density is an obvious indicator of reactivity of radicals (in which there is a single unpaired
electron). Bonds will be made to centers for which the spin density is greatest. For example, the
spin density isosurface for allyl radical suggests that reaction will occur on one of the terminal
carbons and not on the central carbon.
Electrostatic potential
The value of the electrostatic potential (the energy of interaction of a positive point charge with the nuclei
and electrons of a molecule) mapped onto an electron density isosurface may be employed to distinguish
regions on the surface which are electron rich (“basic” or subject to electrophilic attack) from those which
are electron poor (“acidic” or subject to nucleophilic attack).
• Negative potential surfaces serve to “outline” the location of the highest-energy electrons, for example
lone pairs.
Example: A surface for which the electrostatic potential is negative, above and below the plane of
the ring in benzene, and in the ring plane above the nitrogen in pyridine
benzene pyridine
While these two molecules are structurally very similar, potential surfaces make clear that this
similarity does not carry over into their electrophilic reactivities.
Polarization potential
The polarization potential provides the energy due to electronic reorganization of the molecule as a result of
its interaction with a point positive charge. For example, It properly orders the proton affinities (measure of
gas-phase basicity, or energy released when molecule accept a proton) of trimethylamine, dimethyl ether
and fluoromethane.
Local Ionization potential
The local ionization potential is intended to reflect the relative ease of electron removal (“ionization”) at any
location around a molecule. For example, a surface of “low” local ionization potential for sulfur tetrafluoride
demarks the areas which are most easily ionized.
Semiempirical methods
• Semiempirical methods of quantum chemistry start out from the ab initio formalism (HF-SCF) and then
introduce assumptions to speed up the calculations, typically neglegting many of the less important
terms in the ab initio equations.
• In order to compensate for the errors caused by these approximations, empirical parameters are
incorporated into the formalism and calibrated against reliable experimental or theoretical reference
data.
• It is generally recognized that ab initio methods (MP, CI, and CC) and even DFT can give the right result for
the right reason, not only in principle, but often in practice, and that semiempirical calculations can offer
qualitatively correct results of useful accuracy for many larger and chemically interesting systems.
• Semiempirical calculations are usually faster than DFT computations by more that two orders of
magnitude, and therefore they often remain the method of choice in applications that involve really large
molecules (biochemistry) or a large number of molecules or a large number of calculations (dynamics).
• Today, many chemical problems are solved by the combined use of ab initio, DFT, and semiempirical
methods.
Basic concepts:
• A semiempirical model employs a Hartree-Fock SCF-MO treatment for the valence electrons with a
minimal basis set.
• The core electrons are taken into a account through the effective nuclear charge, which is used in
place of the actual nuclear charge to account for the electron-electron repulsions, or represented
by ECP.
• Dynamic electron correlation effects are often included in an average sense.
• The standard HF equations are simplified by neglegting all three-center and four-center two
electron integrals.
• One-center and two-center integral calculations are also simplified. For example CNDO (complete
neglect of differential overlap), INDO (intermediate neglect of differential overlap), and NDDO
(neglect of diatomic differential overlap) schemes differ how they introduce approximations in one-
center and two-center integral calculations.
Consider the following two-electron integral
1
𝜓𝑖∗ 1 𝜓𝑗∗ 2 𝜓 2 𝜓𝑖 1 𝑑𝜏1 𝑑𝜏2
𝑟12 𝑗
𝜓𝑖 = 𝑐𝑖𝑘 𝜙𝑘
𝑘
The zero differential overlap approximation ignores integrals that contain the products where 𝑘 is not
equal to 𝑙 and 𝑟 is not equal to 𝑠
𝑘𝑙 𝑟𝑠 = 𝑘𝑘 𝑟𝑟
total number of such integrals is reduced approximately from 𝑁 4 8 (Hartree Fock) to 𝑁 2 2.
• The CNDO method use the zero differential overlap approximation completely.
Spherically symmetric orbitals only
• Methods based on the intermediate neglect of differential overlap, such as INDO, do not apply the
zero differential overlap approximation when all four basis functions are on the same atom
One-centre repulsion integrals between different orbitals
• Methods that use the neglect of diatomic differential overlap, NDDO, do not apply the zero
differential overlap approximation when the basis functions for the first electron are on the same
atom and the basis functions for the second electron are on the same atom.
Includes some directionality of orbitals
• The approximations work reasonably well when the integrals that remain are parametrized
• The one-center and two-center integrals are determined directly from experimental data (one-
center integrals derived from atomic spectroscopic data) or calculated using analytical formulas or
represented by suitable parametric expressions (empirical or high-level ab initio).
• Most succesful semi-empirical methods (for studying ground-state potential energy surfaces) are
based on NDDO scheme
o In MNDO (Modified Neglect of Diatomic Overlap), the parametrisation is focused on ground
state properties (heats of formation and geometries), ionization potentials, and dipole
moments
o Later, MNDO was essentially replaced by two new methods, PM3 and AM1, which are similar
but have different parametrisation methods (more parameters and thus, more flexibility).
Applications:
• Large biomolecules with thousands of atoms
o The accuracy of semiempirical methods is best for organic compounds
• Medicinal chemistry and drug design
o Semiempirical methods are well suited for quantitative structure-property relationship and
quantitative structure-activity relationship (QSPR and QSAR, respectively) modeling.
• Nanoparticles
o Large fullerenes and nanotubes are prime examples
• Solids and surfaces
o Large clusters which approach bulk limit
• Direct reaction dynamics
o Thousands or even millions single point calculations
• Electronically excited states of large molecules and photochemistry. Alternative for TDDFT.
Molecular dynamics
• Molecular dynamics (MD) is computer simulation technique where the time evolution
of atoms is followed by solving their equations of motions.
• It uses a Maxwell-Botzmann averaging for thermodynamic properties
• Results emerge in a form of simulation. Changes in structures of systems, vibrations, as
well as movements of particles are simulated.
• Simulation ”brings to life” the models yielding vast array of chemical and physical
information often surpassing (in content at least) the real experiments.
Molecular dynamics basics
• The laws of classical mechanics are followed, most notably Newton’s law:
F𝑖 = 𝑚𝑖 a𝑖
for each atom 𝑖 in a system constituted by N atoms. Here, 𝑚𝑖 is the atom mass, a𝑖 = 𝑑2 r𝑖 𝑑𝑡 2 its
acceleration, and F𝑖 the force acting upon it.
• MD is a deterministic technique: given an initial set of positions and velocities, the subsequent
time evolution is in principle completely determined.
• In practice small numerical errors cause chaotic behaviour (butterfly effect).
• MD is a statistical mechanics method. It is a way to obtain a set of configurations or states
distributed according to some statistical distribution function, or statistical ensemble.
• The properties, such as kinetic energy for example, are calculated using time averages.
These are assumed to correspond to observable ensemble averages when the system is
allowed to evolve in time indefinitely so system will eventually pass through all possible
states (Ergodic hypothesis). Because the simulations are of fixed duration, one must be
certain to sample a sufficient amount of phase space.
Phase Space
• For a system of N particles (e.g. atoms), the phase space is the 6N dimensional space of all the positions and
momenta.
• At any given time, the state of the system (i.e. generally, the position and velocity of every atom) is given by
a unique point in the phase space.
• The time evolution of the system can be seen as a displacement of the point in the phase space. Molecular
dynamics as a simulation method is mainly a way of exploring, or sampling, the phase space.
• One of the biggest problem in molecular simulations is that the volume of the phase space (i.e., the number
of accessible configurations for the system) is usually so huge that it is impossible to examine all of it.
• However, (for the case of a constant temperature system, usual in MD) different regions of the phase space
have different probabilities to be observed.
• Boltzmann distribution says that the system has a higher probability to be in a low energy state. Molecular
dynamics can be viewed as a way of producing configurations of the system (so, points in the phase space)
according to their Boltzmann weight.
Modeling the system
• Choosing the potential energy function 𝑉(r1 , … , r𝑁 )
• Deriving the forces as the gradients of the potential with respect to atomic displacements:
F𝑖 = −∇r𝑖 𝑉(r1 , … , r𝑁 )
• Writing the potential as a sum of pairwise interactions:
𝑉 r1 , … , r𝑁 = 𝜙 r𝑖 − r𝑗
𝑖 𝑗>𝑖
𝜙LJ 𝑟 − 𝜙LJ 𝑅𝑐 if 𝑟 ≤ 𝑅𝑐
𝑉 𝑟 =
0 if 𝑟 > 𝑅𝑐
Time integration
• The trajectories of interacting particles are calculated by integrating their equation of motion over time
• Time integration is based on finite difference methods, where time is discretized on a finite grid, the time
step ∆𝑡 being the distance between consecutive points in the grid.
• Knowing the positions and time derivatives at time 𝑡, the integration gives new quantities at a later time
𝑡 + ∆𝑡.
• By iterating the procedure, the time evolution of the system can be followed for long times.
• The most commonly used time integration algorithm is propably the velocity Verlet algorithm, where
position, velocities and accelerations at time 𝑡 + ∆𝑡 are obtained from the same quantities at time 𝑡 in the
following way
𝐫 𝑡 + ∆𝑡 = 𝐫 𝑡 + 𝐯 𝑡 ∆𝑡 + 1 2 𝐚 𝑡 ∆𝑡 2
𝐯 𝑡 + ∆𝑡/2 = 𝐯 𝑡 + 1 2 𝐚 𝑡 ∆𝑡
𝐚 𝑡 + ∆𝑡 = − 1 𝑚 ∇𝑉(𝐫(𝑡 + ∆𝑡))
𝐯 𝑡 + ∆𝑡 = 𝐯 𝑡 + ∆𝑡/2 + 1 2 𝐚 𝑡 + ∆𝑡 ∆𝑡
• Velocities are required (to obtain kinetic energy 𝐾) to test the conservation of energy 𝐸 = 𝐾 + 𝑉.
Application Areas for MD
A proton transfer dynamics of 2-aminopyridine dimer studied using ab initio molecular dynamics (QM-MD)
(Phys. Chem. Chem. Phys., 2011, 13, 5881-5887)
Examples:
Glycine
Limitations of molecular dynamics
• Electrons are not present explicitly in classical MD
• Realism of potential energy surfaces
o Parameters are imperfect
• Classical description of atomic motions
o Quantum effects can become significant in any system as soon as T is sufficiently low
• The limitations on the size (number of atoms) and time of the simulation constrain the range of
problems that can be addressed by the MD method
o The size of structural features of interest are limited to the size of the MD computational cell
on the order of tens of nm.
o Using modern computers it is possible to calculate 106– 108 timesteps. Therefore MD can
only simulate processes that occur within 1 – 100 ns. This is a serious limitation for many
problems, for example thermally-activated processes.
o Record: Largest MD simulation
Periodic boundary conditions
• Particles are enclosed in a box
• Box is replicated to infinity by rigid translation to the
cartesian directions
• All these replicated particles move together but only one of
them is represented in a computer program
• Each particle in the box is not only interacting with other
particles in the same box, but also with their images in
other boxes
• In practice the number of interactions is limited because
cutoff radius of interaction potential
• Surface can be modeled by creating a slab, having periodic
boundary conditions only in two directions
Running a simulation
• Early MD simulations were performed in the NVE ensemble, recent methods perform also at NVT (and
NPT)
• During a simulation at constant energy the temperature will fluctuate. The temperature can be calculated
from the velocities using the relationship:
3
𝐸𝑘 = 𝑁𝑘𝐵 𝑇
2
𝑁
1
𝑇= 𝑚𝑖 𝑣𝑖2
3𝑁𝑘𝐵
𝑖=1
Constant Temperature:
• Couple the system to an external heat bath that is fixed at the desired temperature
• The bath acts as a source of thermal energy, supplying or removing heat from the system as needed
Three essential steps taken to set up and run MD simulation:
1. Initialization
o Specify the initial coordinates and velocities
o Velocities are chosen randomly from Maxwell-Boltzmann or Gaussian distribution
2. Equilibrium
o Prior to starting a simulation it is advisable to do energy minimization
o Useful in correcting flaws such as steric clashes between atoms and distorted bond angles/lengths
3. Production
o Calculate thermodynamic properties of the system
MD simulations examples
http://www2.chem.uic.edu/pkral/movies.html
http://www.ks.uiuc.edu/Gallery/Movies/ProteinsBindingMolecules/
http://www.biochem-caflisch.uzh.ch/movie/13/









