78 |
|
\end{equation} |
79 |
|
|
80 |
|
The solution to Eq.~\ref{introEq:SM2} maximizes the number of |
81 |
< |
degenerative configurations in $E$. \cite{Frenkel1996} |
81 |
> |
degenerate configurations in $E$. \cite{Frenkel1996} |
82 |
|
This gives |
83 |
|
\begin{equation} |
84 |
|
\frac{\partial \ln \Omega(E)}{\partial E_{\gamma}} = 0 = |
197 |
|
|
198 |
|
\subsection{\label{introSec:ergodic}The Ergodic Hypothesis} |
199 |
|
|
200 |
< |
One last important consideration is that of ergodicity. Ergodicity is |
201 |
< |
the assumption that given an infinite amount of time, a system will |
202 |
< |
visit every available point in phase space.\cite{Frenkel1996} For most |
203 |
< |
systems, this is a valid assumption, except in cases where the system |
204 |
< |
may be trapped in a local feature (\emph{e.g.}~glasses). When valid, |
205 |
< |
ergodicity allows the unification of a time averaged observation and |
206 |
< |
an ensemble averaged one. If an observation is averaged over a |
207 |
< |
sufficiently long time, the system is assumed to visit all |
208 |
< |
appropriately available points in phase space, giving a properly |
209 |
< |
weighted statistical average. This allows the researcher freedom of |
210 |
< |
choice when deciding how best to measure a given observable. When an |
211 |
< |
ensemble averaged approach seems most logical, the Monte Carlo |
212 |
< |
techniques described in Sec.~\ref{introSec:monteCarlo} can be utilized. |
213 |
< |
Conversely, if a problem lends itself to a time averaging approach, |
214 |
< |
the Molecular Dynamics techniques in Sec.~\ref{introSec:MD} can be |
215 |
< |
employed. |
200 |
> |
In the case of a Molecular Dynamics simulation, rather than |
201 |
> |
calculating an ensemble average integral over phase space as in |
202 |
> |
Eq.~\ref{introEq:SM15}, it becomes easier to caclulate the time |
203 |
> |
average of an observable. Namely, |
204 |
> |
\begin{equation} |
205 |
> |
\langle A \rangle_t = \frac{1}{\tau} |
206 |
> |
\int_0^{\tau} A[\boldsymbol{\Gamma}(t)]\,dt |
207 |
> |
\label{introEq:SM16} |
208 |
> |
\end{equation} |
209 |
> |
Where the value of an observable is averaged over the length of time |
210 |
> |
that the simulation is run. This type of measurement mirrors the |
211 |
> |
experimental measurement of an observable. In an experiment, the |
212 |
> |
instrument analyzing the system must average its observation of the |
213 |
> |
finite time of the measurement. What is required then, is a principle |
214 |
> |
to relate the time average to the ensemble average. This is the |
215 |
> |
ergodic hypothesis. |
216 |
|
|
217 |
+ |
Ergodicity is the assumption that given an infinite amount of time, a |
218 |
+ |
system will visit every available point in phase |
219 |
+ |
space.\cite{Frenkel1996} For most systems, this is a valid assumption, |
220 |
+ |
except in cases where the system may be trapped in a local feature |
221 |
+ |
(\emph{e.g.}~glasses). When valid, ergodicity allows the unification |
222 |
+ |
of a time averaged observation and an ensemble averaged one. If an |
223 |
+ |
observation is averaged over a sufficiently long time, the system is |
224 |
+ |
assumed to visit all appropriately available points in phase space, |
225 |
+ |
giving a properly weighted statistical average. This allows the |
226 |
+ |
researcher freedom of choice when deciding how best to measure a given |
227 |
+ |
observable. When an ensemble averaged approach seems most logical, |
228 |
+ |
the Monte Carlo techniques described in Sec.~\ref{introSec:monteCarlo} |
229 |
+ |
can be utilized. Conversely, if a problem lends itself to a time |
230 |
+ |
averaging approach, the Molecular Dynamics techniques in |
231 |
+ |
Sec.~\ref{introSec:MD} can be employed. |
232 |
+ |
|
233 |
|
\section{\label{introSec:monteCarlo}Monte Carlo Simulations} |
234 |
|
|
235 |
|
The Monte Carlo method was developed by Metropolis and Ulam for their |
247 |
|
\end{equation} |
248 |
|
The equation can be recast as: |
249 |
|
\begin{equation} |
250 |
< |
I = (b-a)\langle f(x) \rangle |
250 |
> |
I = \int^b_a \frac{f(x)}{\rho(x)} \rho(x) dx |
251 |
> |
\label{introEq:Importance1} |
252 |
> |
\end{equation} |
253 |
> |
Where $\rho(x)$ is an arbitrary probability distribution in $x$. If |
254 |
> |
one conducts $\tau$ trials selecting a random number, $\zeta_\tau$, |
255 |
> |
from the distribution $\rho(x)$ on the interval $[a,b]$, then |
256 |
> |
Eq.~\ref{introEq:Importance1} becomes |
257 |
> |
\begin{equation} |
258 |
> |
I= \lim_{\tau \rightarrow \infty}\biggl \langle \frac{f(x)}{\rho(x)} \biggr \rangle_{\text{trials}[a,b]} |
259 |
> |
\label{introEq:Importance2} |
260 |
> |
\end{equation} |
261 |
> |
If $\rho(x)$ is uniformly distributed over the interval $[a,b]$, |
262 |
> |
\begin{equation} |
263 |
> |
\rho(x) = \frac{1}{b-a} |
264 |
> |
\label{introEq:importance2b} |
265 |
> |
\end{equation} |
266 |
> |
then the integral can be rewritten as |
267 |
> |
\begin{equation} |
268 |
> |
I = (b-a)\lim_{\tau \rightarrow \infty} |
269 |
> |
\langle f(x) \rangle_{\text{trials}[a,b]} |
270 |
|
\label{eq:MCex2} |
271 |
|
\end{equation} |
237 |
– |
Where $\langle f(x) \rangle$ is the unweighted average over the interval |
238 |
– |
$[a,b]$. The calculation of the integral could then be solved by |
239 |
– |
randomly choosing points along the interval $[a,b]$ and calculating |
240 |
– |
the value of $f(x)$ at each point. The accumulated average would then |
241 |
– |
approach $I$ in the limit where the number of trials is infinitely |
242 |
– |
large. |
272 |
|
|
273 |
|
However, in Statistical Mechanics, one is typically interested in |
274 |
|
integrals of the form: |
290 |
|
average. This is where importance sampling comes into |
291 |
|
play.\cite{allen87:csl} |
292 |
|
|
293 |
< |
Importance Sampling is a method where one selects a distribution from |
294 |
< |
which the random configurations are chosen in order to more |
295 |
< |
efficiently calculate the integral.\cite{Frenkel1996} Consider again |
296 |
< |
Eq.~\ref{eq:MCex1} rewritten to be: |
268 |
< |
\begin{equation} |
269 |
< |
I = \int^b_a \frac{f(x)}{\rho(x)} \rho(x) dx |
270 |
< |
\label{introEq:Importance1} |
271 |
< |
\end{equation} |
272 |
< |
Where $\rho(x)$ is an arbitrary probability distribution in $x$. If |
273 |
< |
one conducts $\tau$ trials selecting a random number, $\zeta_\tau$, |
274 |
< |
from the distribution $\rho(x)$ on the interval $[a,b]$, then |
275 |
< |
Eq.~\ref{introEq:Importance1} becomes |
276 |
< |
\begin{equation} |
277 |
< |
I= \biggl \langle \frac{f(x)}{\rho(x)} \biggr \rangle_{\text{trials}} |
278 |
< |
\label{introEq:Importance2} |
279 |
< |
\end{equation} |
280 |
< |
Looking at Eq.~\ref{eq:mcEnsAvg}, and realizing |
293 |
> |
Importance sampling is a method where the distribution, from which the |
294 |
> |
random configurations are chosen, is selected in such a way as to |
295 |
> |
efficiently sample the integral in question. Looking at |
296 |
> |
Eq.~\ref{eq:mcEnsAvg}, and realizing |
297 |
|
\begin {equation} |
298 |
|
\rho_{kT}(\mathbf{r}^N) = |
299 |
|
\frac{e^{-\beta V(\mathbf{r}^N)}} |
317 |
|
By selecting $\rho(\mathbf{r}^N)$ to be $\rho_{kT}(\mathbf{r}^N)$ |
318 |
|
Eq.~\ref{introEq:Importance4} becomes |
319 |
|
\begin{equation} |
320 |
< |
\langle A \rangle = \langle A(\mathbf{r}^N) \rangle_{\text{trials}} |
320 |
> |
\langle A \rangle = \langle A(\mathbf{r}^N) \rangle_{kT} |
321 |
|
\label{introEq:Importance5} |
322 |
|
\end{equation} |
323 |
|
The difficulty is selecting points $\mathbf{r}^N$ such that they are |
435 |
|
integrated in order to obtain information about both the positions and |
436 |
|
momentum of a system, allowing the calculation of not only |
437 |
|
configurational observables, but momenta dependent ones as well: |
438 |
< |
diffusion constants, velocity auto correlations, folding/unfolding |
439 |
< |
events, etc. Due to the principle of ergodicity, |
438 |
> |
diffusion constants, relaxation events, folding/unfolding |
439 |
> |
events, etc. With the time dependent information gained from a |
440 |
> |
Molecular Dynamics simulation, one can also calculate time correlation |
441 |
> |
functions of the form\cite{Hansen86} |
442 |
> |
\begin{equation} |
443 |
> |
\langle A(t)\,A(0)\rangle = \lim_{\tau\rightarrow\infty} \frac{1}{\tau} |
444 |
> |
\int_0^{\tau} A(t+t^{\prime})\,A(t^{\prime})\,dt^{\prime} |
445 |
> |
\label{introEq:timeCorr} |
446 |
> |
\end{equation} |
447 |
> |
These correlations can be used to measure fundamental time constants |
448 |
> |
of a system, such as diffusion constants from the velocity |
449 |
> |
autocorrelation or dipole relaxation times from the dipole |
450 |
> |
autocorrelation. Due to the principle of ergodicity, |
451 |
|
Sec.~\ref{introSec:ergodic}, the average of these observables over the |
452 |
|
time period of the simulation are taken to be the ensemble averages |
453 |
|
for the system. |
454 |
|
|
455 |
|
The choice of when to use molecular dynamics over Monte Carlo |
456 |
|
techniques, is normally decided by the observables in which the |
457 |
< |
researcher is interested. If the observables depend on momenta in |
457 |
> |
researcher is interested. If the observables depend on time in |
458 |
|
any fashion, then the only choice is molecular dynamics in some form. |
459 |
|
However, when the observable is dependent only on the configuration, |
460 |
< |
then most of the time Monte Carlo techniques will be more efficient. |
460 |
> |
then for most small systems, Monte Carlo techniques will be more efficient. |
461 |
|
|
462 |
|
The focus of research in the second half of this dissertation is |
463 |
|
centered around the dynamic properties of phospholipid bilayers, |
547 |
|
efficiency of the force evaluation, as particles farther than a |
548 |
|
predetermined distance, $r_{\text{cut}}$, are not included in the |
549 |
|
calculation.\cite{Frenkel1996} In a simulation with periodic images, |
550 |
< |
$r_{\text{cut}}$ has a maximum value of $\text{box}/2$. |
551 |
< |
Fig.~\ref{introFig:rMax} illustrates how when using an |
552 |
< |
$r_{\text{cut}}$ larger than this value, or in the extreme limit of no |
553 |
< |
$r_{\text{cut}}$ at all, the corners of the simulation box are |
554 |
< |
unequally weighted due to the lack of particle images in the $x$, $y$, |
555 |
< |
or $z$ directions past a distance of $\text{box} / 2$. |
550 |
> |
there are two methods to choose from, both with their own cutoff |
551 |
> |
limits. In the minimum image convention, $r_{\text{cut}}$ has a |
552 |
> |
maximum value of $\text{box}/2$. This is because each atom has only |
553 |
> |
one image that is seen by another atom, and further the image used is |
554 |
> |
the one that minimizes the distance between the two atoms. A system of |
555 |
> |
wrapped images about a central atom therefore has a maximum length |
556 |
> |
scale of box on a side (Fig.~\ref{introFig:rMaxMin}). The second |
557 |
> |
convention, multiple image convention, has a maximum $r_{\text{cut}}$ |
558 |
> |
of box. Here multiple images of each atom are replicated in the |
559 |
> |
periodic cells surrounding the central atom, this causes the atom to |
560 |
> |
see multiple copies of several atoms. If the cutoff radius is larger |
561 |
> |
than box, however, then the atom will see an image of itself, and |
562 |
> |
attempt to calculate an unphysical self-self force interaction |
563 |
> |
(Fig.~\ref{introFig:rMaxMult}). Due to the increased complexity and |
564 |
> |
commputaional ineffeciency of the multiple image method, the minimum |
565 |
> |
image method is the periodic method used throughout this research. |
566 |
|
|
567 |
|
\begin{figure} |
568 |
|
\centering |
569 |
|
\includegraphics[width=\linewidth]{rCutMaxFig.eps} |
570 |
< |
\caption[An explanation of $r_{\text{cut}}$]{The yellow atom has all other images wrapped to itself as the center. If $r_{\text{cut}}=\text{box}/2$, then the distribution is uniform (blue atoms). However, when $r_{\text{cut}}>\text{box}/2$ the corners are disproportionately weighted (green atoms) vs the axial directions (shaded regions).} |
571 |
< |
\label{introFig:rMax} |
570 |
> |
\caption[An explanation of minimum image convention]{The yellow atom has all other images wrapped to itself as the center. If $r_{\text{cut}}=\text{box}/2$, then the distribution is uniform (blue atoms). However, when $r_{\text{cut}}>\text{box}/2$ the corners are disproportionately weighted (green atoms) vs the axial directions (shaded regions).} |
571 |
> |
\label{introFig:rMaxMin} |
572 |
> |
\end{figure} |
573 |
> |
|
574 |
> |
\begin{figure} |
575 |
> |
\centering |
576 |
> |
\includegraphics[width=\linewidth]{rCutMaxMultFig.eps} |
577 |
> |
\caption[An explanation of multiple image convention]{The yellow atom is the central wrapping point. The blue atoms are the minimum images of the system about the central atom. The boxes with the green atoms are multiple images of the central box. If $r_{\text{cut}} \geq \{text{box}$ then the central atom sees multiple images of itself (red atom), creating a self-self force evaluation.} |
578 |
> |
\label{introFig:rMaxMult} |
579 |
|
\end{figure} |
580 |
|
|
581 |
|
With the use of an $r_{\text{cut}}$, however, comes a discontinuity in |