ViewVC Help
View File | Revision Log | Show Annotations | View Changeset | Root Listing
root/group/trunk/xDissertation/Introduction.tex
(Generate patch)

Comparing trunk/xDissertation/Introduction.tex (file contents):
Revision 3374 by xsun, Thu Mar 20 22:21:43 2008 UTC vs.
Revision 3375 by xsun, Mon Mar 24 20:12:52 2008 UTC

# Line 3 | Line 3 | because of their critical role as the foundation of bi
3   \section{Background on the Problem\label{In:sec:pro}}
4   Phospholipid molecules are the primary topic of this dissertation
5   because of their critical role as the foundation of biological
6 < membranes. The chemical structure of phospholipids includes the polar
7 < head group which is due to a large charge separation between phosphate
8 < and amino alcohol, and the nonpolar tails that contains fatty acid
9 < chains. Depending on the alcohol which phosphate and fatty acid chains
10 < are esterified to, the phospholipids are divided into
11 < glycerophospholipids and sphingophospholipids.~\cite{Cevc80} The
12 < chemical structures are shown in figure~\ref{Infig:lipid}.
6 > membranes. The chemical structure of phospholipids includes a head
7 > group with a large dipole moment which is due to the large charge
8 > separation between phosphate and amino alcohol, and a nonpolar tail
9 > that contains fatty acid chains. Depending on the specific alcohol
10 > which the phosphate and fatty acid chains are esterified to, the
11 > phospholipids are divided into glycerophospholipids and
12 > sphingophospholipids.~\cite{Cevc80} The chemical structures are shown
13 > in figure~\ref{Infig:lipid}.
14   \begin{figure}
15   \centering
16   \includegraphics[width=\linewidth]{./figures/inLipid.pdf}
# Line 17 | Line 18 | sphingophospholipids (right).\cite{Cevc80}}
18   sphingophospholipids (right).\cite{Cevc80}}
19   \label{Infig:lipid}
20   \end{figure}
21 < The glycerophospholipid is the dominant phospholipid in membranes. The
22 < types of glycerophospholipids depend on the X group, and the
23 < chains. For example, if X is choline, the lipids are known as
24 < phosphatidylcholine (PC), or if X is ethanolamine, the lipids are
25 < known as phosphatidyethanolamine (PE). Table~\ref{Intab:pc} listed a
26 < number types of phosphatidycholine with different fatty acids as the
27 < lipid chains.
21 > Glycerophospholipids are the dominant phospholipids in biological
22 > membranes. The type of glycerophospholipid depends on the identity of
23 > the X group, and the chains. For example, if X is choline
24 > [(CH$_3$)$_3$N$^+$CH$_2$CH$_2$OH], the lipids are known as
25 > phosphatidylcholine (PC), or if X is ethanolamine
26 > [H$_3$N$^+$CH$_2$CH$_2$OH], the lipids are known as
27 > phosphatidyethanolamine (PE). Table~\ref{Intab:pc} listed a number
28 > types of phosphatidycholine with different fatty acids as the lipid
29 > chains.
30   \begin{table*}
31   \begin{minipage}{\linewidth}
32   \begin{center}
# Line 48 | Line 51 | various phases has not been achieved.
51   complete understanding of the mechanism and driving forces behind the
52   various phases has not been achieved.
53  
54 < \subsection{Ripple Phase\label{In:ssec:ripple}}
54 > \subsection{The Ripple Phase\label{In:ssec:ripple}}
55   The $P_{\beta'}$ {\it ripple phase} of lipid bilayers, named from the
56   periodic buckling of the membrane, is an intermediate phase which is
57   developed either from heating the gel phase $L_{\beta'}$ or cooling
58 < the fluid phase $L_\alpha$. A Sketch is shown in
58 > the fluid phase $L_\alpha$. A sketch of the phases is shown in
59   figure~\ref{Infig:phaseDiagram}.~\cite{Cevc80}
60   \begin{figure}
61   \centering
62   \includegraphics[width=\linewidth]{./figures/inPhaseDiagram.pdf}
63 < \caption{A phase diagram of lipid bilayer. With increasing the
64 < temperature, the bilayer can go through a gel ($L_{\beta'}$), ripple
65 < ($P_{\beta'}$) to fluid ($L_\alpha$) phase transition.~\cite{Cevc80}}
63 > \caption{Phases of PC lipid bilayers. With increasing
64 > temperature, phosphotidylcholine (PC) bilayers can go through
65 > $L_{\beta'} \rightarrow P_{\beta'}$ (gel $\rightarrow$ ripple) and
66 > $P_{\beta'} \rightarrow L_\alpha$ (ripple $\rightarrow$ fluid) phase
67 > transitions.~\cite{Cevc80}}
68   \label{Infig:phaseDiagram}
69   \end{figure}
70 < Most structural information of the ripple phase has been obtained by
71 < the X-ray diffraction~\cite{Sun96,Katsaras00} and freeze-fracture
70 > Most structural information about the ripple phase has been obtained
71 > by X-ray diffraction~\cite{Sun96,Katsaras00} and freeze-fracture
72   electron microscopy (FFEM).~\cite{Copeland80,Meyer96} The X-ray
73   diffraction work by Katsaras {\it et al.} showed that a rich phase
74   diagram exhibiting both {\it asymmetric} and {\it symmetric} ripples
# Line 74 | Line 79 | mica.~\cite{Kaasgaard03}
79   \begin{figure}
80   \centering
81   \includegraphics[width=\linewidth]{./figures/inRipple.pdf}
82 < \caption{The experimental observed ripple phase. The top image is
83 < obtained by Sun {\it et al.} using X-ray diffraction~\cite{Sun96},
84 < and the bottom one is observed by Kaasgaard {\it et al.} using
85 < AFM.~\cite{Kaasgaard03}}
82 > \caption{Experimental observations of the riple phase. The top image
83 > is an electrostatic density map obtained by Sun {\it et al.} using
84 > X-ray diffraction~\cite{Sun96}.  The lower figures are the surface
85 > topology of various ripple domains in bilayers supported in mica. The
86 > AFM images were observed by Kaasgaard {\it et
87 > al.}.~\cite{Kaasgaard03}}
88   \label{Infig:ripple}
89   \end{figure}
90 < Figure~\ref{Infig:ripple} shows the ripple phase oberved by X-ray
90 > Figure~\ref{Infig:ripple} shows the ripple phase oberved by both X-ray
91   diffraction and AFM. The experimental results provide strong support
92   for a 2-dimensional triangular packing lattice of the lipid molecules
93   within the ripple phase.  This is a notable change from the observed
94   lipid packing within the gel phase,~\cite{Cevc80} although Tenchov
95 < {\it et al.} have recently observed near-hexagonal packing in some
95 > {\it et al.} have recently observed near-triangular packing in some
96   phosphatidylcholine (PC) gel phases.~\cite{Tenchov2001} However, the
97   physical mechanism for the formation of the ripple phase has never
98   been explained and the microscopic structure of the ripple phase has
99   never been elucidated by experiments. Computational simulation is a
100   perfect tool to study the microscopic properties for a
101 < system. However, the large length scale the ripple structure and the
102 < long time scale of the formation of the ripples are crucial obstacles
103 < to performing the actual work. The principal ideas explored in this
104 < dissertation are attempts to break the computational task up by
101 > system. However, the large length scale of the ripple structures and
102 > the long time required for the formation of the ripples are crucial
103 > obstacles to performing the actual work. The principal ideas explored
104 > in this dissertation are attempts to break the computational task up
105 > by
106   \begin{itemize}
107   \item Simplifying the lipid model.
108 < \item Improving algorithm for integrating the equations of motion.
108 > \item Improving the algorithm for integrating the equations of motion.
109   \end{itemize}
110   In chapters~\ref{chap:mc} and~\ref{chap:md}, we use a simple point
111 < dipole spin model and a coarse-grained molecualr scale model to
111 > dipole spin model and a coarse-grained molecular scale model to
112   perform the Monte Carlo and Molecular Dynamics simulations
113   respectively, and in chapter~\ref{chap:ld}, we develop a Langevin
114   Dynamics algorithm which excludes the explicit solvent to improve the
115   efficiency of the simulations.
116  
117 < \subsection{Lattice Model\label{In:ssec:model}}
118 < The gel-like characteristic (relatively immobile molecules) exhibited
117 > \subsection{Lattice Models\label{In:ssec:model}}
118 > The gel-like characteristic (laterally immobile molecules) exhibited
119   by the ripple phase makes it feasible to apply a lattice model to
120   study the system. The popular $2$ dimensional lattice models, {\it
121   i.e.}, the Ising, $X-Y$, and Heisenberg models, show {\it frustration}
# Line 130 | Line 138 | represented by arrows. The multiple local minima of en
138   \includegraphics[width=3in]{./figures/inFrustration.pdf}
139   \caption{Frustration on triangular lattice, the spins and dipoles are
140   represented by arrows. The multiple local minima of energy states
141 < induce the frustration for spins and dipoles picking the directions.}
141 > induce frustration for spins and dipoles resulting in disordered
142 > low-temperature phases.}
143   \label{Infig:frustration}
144   \end{figure}
145 < The spins in figure~\ref{Infig:frustration} shows an illustration of
146 < the frustration for $J < 0$ on a triangular lattice. There are
147 < multiple local minima energy states which are independent of the
148 < direction of the spin on top of the triangle, therefore infinite
149 < possibilities for the packing of spins which induces what is known as
150 < ``complete regular frustration'' which leads to disordered low
151 < temperature phases. The similarity goes to the dipoles on a hexagonal
152 < lattice, which are shown by the dipoles in
153 < figure~\ref{Infig:frustration}. In this circumstance, the dipoles want
154 < to be aligned, however, due to the long wave fluctuation, at low
155 < temperature, the aligned state becomes unstable, vortex is formed and
156 < results in multiple local minima of energy states. The dipole on the
148 < center of the hexagonal lattice is frustrated.
145 > The spins in figure~\ref{Infig:frustration} illustrate frustration for
146 > $J < 0$ on a triangular lattice. There are multiple local minima
147 > energy states which are independent of the direction of the spin on
148 > top of the triangle, therefore infinite possibilities for orienting
149 > large numbers spins. This induces what is known as ``complete regular
150 > frustration'' which leads to disordered low temperature phases. This
151 > behavior extends to dipoles on a triangular lattices, which are shown
152 > in the lower portion of figure~\ref{Infig:frustration}. In this case,
153 > dipole-aligned structures are energetically favorable, however, at low
154 > temperature, vortices are easily formed, and, this results in multiple
155 > local minima of energy states for a central dipole. The dipole on the
156 > center of the hexagonal lattice is therefore frustrated.
157  
158   The lack of translational degrees of freedom in lattice models
159   prevents their utilization in models for surface buckling. In
# Line 159 | Line 167 | principles of statistical mechanics.~\cite{Tolman1979}
167   to key concepts of classical statistical mechanics that we used in
168   this dissertation. Tolman gives an excellent introduction to the
169   principles of statistical mechanics.~\cite{Tolman1979} A large part of
170 < section~\ref{In:sec:SM} will follow Tolman's notation.
170 > section~\ref{In:sec:SM} follows Tolman's notation.
171  
172   \subsection{Ensembles\label{In:ssec:ensemble}}
173   In classical mechanics, the state of the system is completely
# Line 269 | Line 277 | selected moving representative points in the phase spa
277   and the rate of density change is zero in the neighborhood of any
278   selected moving representative points in the phase space.
279  
280 < The condition of the ensemble is determined by the density
280 > The type of thermodynamic ensemble is determined by the density
281   distribution. If we consider the density distribution as only a
282   function of $q$ and $p$, which means the rate of change of the phase
283   space density in the neighborhood of all representative points in the
# Line 298 | Line 306 | constant everywhere in the phase space,
306   \rho = \mathrm{const.}
307   \label{Ineq:uniformEnsemble}
308   \end{equation}
309 < the ensemble is called {\it uniform ensemble}.
309 > the ensemble is called {\it uniform ensemble}, but this ensemble has
310 > little relevance for physical chemistry. It is an ensemble with
311 > essentially infinite temperature.
312  
313   \subsubsection{The Microcanonical Ensemble\label{In:sssec:microcanonical}}
314 < Another useful ensemble is the {\it microcanonical ensemble}, for
315 < which:
314 > The most useful ensemble for Molecular Dynamics is the {\it
315 > microcanonical ensemble}, for which:
316   \begin{equation}
317   \rho = \delta \left( H(q^N, p^N) - E \right) \frac{1}{\Sigma (N, V, E)}
318   \label{Ineq:microcanonicalEnsemble}
# Line 319 | Line 329 | makes the argument of $\ln$ dimensionless. In this cas
329   \end{equation}
330   where $k_B$ is the Boltzmann constant and $C^N$ is a number which
331   makes the argument of $\ln$ dimensionless. In this case, $C^N$ is the
332 < total phase space volume of one state. The entropy of a microcanonical
333 < ensemble is given by
332 > total phase space volume of one state which has the same units as
333 > $\Sigma(N, V, E)$. The entropy of a microcanonical ensemble is given
334 > by
335   \begin{equation}
336   S = k_B \ln \left(\frac{\Sigma(N, V, E)}{C^N}\right).
337   \label{Ineq:entropy}
# Line 337 | Line 348 | Z_N = \int d \vec q~^N \int_\Gamma d \vec p~^N  e^{-H(
348   Z_N = \int d \vec q~^N \int_\Gamma d \vec p~^N  e^{-H(q^N, p^N) / k_B T},
349   \label{Ineq:partitionFunction}
350   \end{equation}
351 < which is also known as the canonical{\it partition function}. $\Gamma$
352 < indicates that the integral is over all phase space. In the canonical
353 < ensemble, $N$, the total number of particles, $V$, total volume, and
354 < $T$, the temperature, are constants. The systems with the lowest
355 < energies hold the largest population. According to maximum principle,
356 < thermodynamics maximizes the entropy $S$, implying that
351 > which is also known as the canonical {\it partition
352 > function}. $\Gamma$ indicates that the integral is over all phase
353 > space. In the canonical ensemble, $N$, the total number of particles,
354 > $V$, total volume, and $T$, the temperature, are constants. The
355 > systems with the lowest energies hold the largest
356 > population. Thermodynamics maximizes the entropy, $S$, implying that
357   \begin{equation}
358   \begin{array}{ccc}
359   \delta S = 0 & \mathrm{and} & \delta^2 S < 0.
# Line 363 | Line 374 | There is an implicit assumption that our arguments are
374   system and the distribution of microscopic states.
375  
376   There is an implicit assumption that our arguments are based on so
377 < far. A representative point in the phase space is equally likely to be
378 < found in any energetically allowed region. In other words, all
379 < energetically accessible states are represented equally, the
380 < probabilities to find the system in any of the accessible states is
381 < equal. This is called the principle of equal a {\it priori}
377 > far. Tow representative points in phase space are equally likely to be
378 > found if they have the same energy. In other words, all energetically
379 > accessible states are represented , and the probabilities to find the
380 > system in any of the accessible states is equal to that states
381 > Boltzmann weight. This is called the principle of equal a {\it priori}
382   probabilities.
383  
384   \subsection{Statistical Averages\label{In:ssec:average}}
# Line 402 | Line 413 | mechanics, so Eq.~\ref{Ineq:timeAverage1} becomes
413   \frac{1}{T} \int_{0}^{T} F[q^N(t), p^N(t)] dt
414   \label{Ineq:timeAverage2}
415   \end{equation}
416 < for an infinite time interval.
416 > for an finite time interval, $T$.
417  
418   \subsubsection{Ergodicity\label{In:sssec:ergodicity}}
419   The {\it ergodic hypothesis}, an important hypothesis governing modern
# Line 445 | Line 456 | C_{AB}(\tau) = \langle A(0)B(\tau) \rangle = \lim_{T \
456   \frac{1}{T} \int_{0}^{T} dt A(t) B(t + \tau),
457   \label{Ineq:crosscorrelationFunction}
458   \end{equation}
459 < and called {\it cross correlation function}.
459 > and is called a {\it cross correlation function}.
460  
461   We know from the ergodic hypothesis that there is a relationship
462   between time average and ensemble average. We can put the correlation
463 < function in a classical mechanics form,
463 > function in a classical mechanical form,
464   \begin{equation}
465 < C_{AA}(\tau) = \int d \vec q~^N \int d \vec p~^N A[(q^N(t), p^N(t)]
466 < A[q^N(t+\tau), p^N(t+\tau)] \rho(q, p)
465 > C_{AA}(\tau) = \int d \vec q~^N \int d \vec p~^N A[(q^N, p^N]
466 > A[q^N(\tau), p^N(\tau)] \rho(q^N, p^N)
467   \label{Ineq:autocorrelationFunctionCM}
468   \end{equation}
469 < and
469 > where $q^N(\tau)$, $p^N(\tau)$ is the phase space point that follows
470 > classical evolution of the point $q^N$, $p^N$ after a tme $\tau$ has
471 > elapsed, and
472   \begin{equation}
473 < C_{AB}(\tau) = \int d \vec q~^N \int d \vec p~^N A[(q^N(t), p^N(t)]
474 < B[q^N(t+\tau), p^N(t+\tau)] \rho(q, p)
473 > C_{AB}(\tau) = \int d \vec q~^N \int d \vec p~^N A[(q^N, p^N]
474 > B[q^N(\tau), p^N(\tau)] \rho(q^N, p^N)
475   \label{Ineq:crosscorrelationFunctionCM}
476   \end{equation}
477   as the autocorrelation function and cross correlation functions
# Line 479 | Line 492 | while Molecular Dynamics simulations provide dynamic
492   simulations. One is that the Monte Carlo simulations are time
493   independent methods of sampling structural features of an ensemble,
494   while Molecular Dynamics simulations provide dynamic
495 < information. Additionally, Monte Carlo methods are a stochastic
496 < processes, the future configurations of the system are not determined
495 > information. Additionally, Monte Carlo methods are stochastic
496 > processes; the future configurations of the system are not determined
497   by its past. However, in Molecular Dynamics, the system is propagated
498 < by Newton's equation of motion, and the trajectory of the system
498 > by Hamilton's equations of motion, and the trajectory of the system
499   evolving in phase space is deterministic. Brief introductions of the
500   two algorithms are given in section~\ref{In:ssec:mc}
501   and~\ref{In:ssec:md}. Langevin Dynamics, an extension of the Molecular
# Line 493 | Line 506 | algorithm is usually applied to the canonical ensemble
506   A Monte Carlo integration algorithm was first introduced by Metropolis
507   {\it et al.}~\cite{Metropolis53} The basic Metropolis Monte Carlo
508   algorithm is usually applied to the canonical ensemble, a
509 < Boltzmann-weighted ensemble, in which the $N$, the total number of
510 < particles, $V$, total volume, $T$, temperature are constants. An
511 < average in this ensemble is given
509 > Boltzmann-weighted ensemble, in which $N$, the total number of
510 > particles, $V$, the total volume, and $T$, the temperature are
511 > constants. An average in this ensemble is given
512   \begin{equation}
513   \langle A \rangle = \frac{1}{Z_N} \int d \vec q~^N \int d \vec p~^N
514   A(q^N, p^N) e^{-H(q^N, p^N) / k_B T}.
# Line 519 | Line 532 | Eq.~\ref{Ineq:configurationIntegral} is equivalent to
532   \rangle$ is now a configuration integral, and
533   Eq.~\ref{Ineq:configurationIntegral} is equivalent to
534   \begin{equation}
535 < \langle A \rangle = \int d \vec q~^N A \rho(q^N).
535 > \langle A \rangle = \int d \vec q~^N A \rho(q^N),
536   \label{Ineq:configurationAve}
537   \end{equation}
538 + where $\rho(q^N)$ is a configurational probability
539 + \begin{equation}
540 + \rho(q^N) = \frac{e^{-U(q^N) / k_B T}}{\int d \vec q~^N e^{-U(q^N) / k_B T}}.
541 + \label{Ineq:configurationProb}
542 + \end{equation}
543  
544   In a Monte Carlo simulation of a system in the canonical ensemble, the
545   probability of the system being in a state $s$ is $\rho_s$, the change
# Line 536 | Line 554 | to state $s'$. Since $\rho_s$ is independent of time a
554   \frac{d \rho_{s}^{equilibrium}}{dt} = 0,
555   \label{Ineq:equiProb}
556   \end{equation}
557 < which means $\sum_{s'} [ -w_{ss'}\rho_s + w_{s's}\rho_{s'} ]$ is $0$
558 < for all $s'$. So
557 > the sum of transition probabilities $\sum_{s'} [ -w_{ss'}\rho_s +
558 > w_{s's}\rho_{s'} ]$ is $0$ for all $s'$. So
559   \begin{equation}
560   \frac{\rho_s^{equilibrium}}{\rho_{s'}^{equilibrium}} = \frac{w_{s's}}{w_{ss'}}.
561   \label{Ineq:relationshipofRhoandW}
# Line 552 | Line 570 | then the ratio of transition probabilities,
570   \frac{w_{s's}}{w_{ss'}} = e^{-(U_s - U_{s'}) / k_B T},
571   \label{Ineq:conditionforBoltzmannStatistics}
572   \end{equation}
573 < An algorithm that shows how Monte Carlo simulation generates a
573 > An algorithm that indicates how a Monte Carlo simulation generates a
574   transition probability governed by
575   \ref{Ineq:conditionforBoltzmannStatistics}, is given schematically as,
576   \begin{enumerate}
577   \item\label{Initm:oldEnergy} Choose a particle randomly, and calculate
578 < the energy of the rest of the system due to the location of the particle.
578 > the energy of the rest of the system due to the current location of
579 > the particle.
580   \item\label{Initm:newEnergy} Make a random displacement of the particle,
581   calculate the new energy taking the movement of the particle into account.
582    \begin{itemize}
583 <     \item If the energy goes down, keep the new configuration and return to
565 < step~\ref{Initm:oldEnergy}.
583 >     \item If the energy goes down, keep the new configuration.
584       \item If the energy goes up, pick a random number between $[0,1]$.
585          \begin{itemize}
586             \item If the random number smaller than
587 < $e^{-(U_{new} - U_{old})} / k_B T$, keep the new configuration and return to
570 < step~\ref{Initm:oldEnergy}.
587 > $e^{-(U_{new} - U_{old})} / k_B T$, keep the new configuration.
588             \item If the random number is larger than
589 < $e^{-(U_{new} - U_{old})} / k_B T$, keep the old configuration and return to
573 < step~\ref{Initm:oldEnergy}.
589 > $e^{-(U_{new} - U_{old})} / k_B T$, keep the old configuration.
590          \end{itemize}
591    \end{itemize}
592   \item\label{Initm:accumulateAvg} Accumulate the averages based on the
593 < current configuartion.
593 > current configuration.
594   \item Go to step~\ref{Initm:oldEnergy}.
595   \end{enumerate}
596 < It is important for sampling accurately that the old configuration is
596 > It is important for sampling accuracy that the old configuration is
597   sampled again if it is kept.
598  
599   \subsection{Molecular Dynamics\label{In:ssec:md}}
# Line 589 | Line 605 | real experiment, the instantaneous positions and momen
605   evolution of the system obeys the laws of classical mechanics, and in
606   most situations, there is no need to account for quantum effects. In a
607   real experiment, the instantaneous positions and momenta of the
608 < particles in the system are neither important nor measurable, the
609 < observable quantities are usually an average value for a finite time
610 < interval. These quantities are expressed as a function of positions
611 < and momenta in Molecular Dynamics simulations. For example,
608 > particles in the system are ofter neither important nor measurable,
609 > the observable quantities are usually an average value for a finite
610 > time interval. These quantities are expressed as a function of
611 > positions and momenta in Molecular Dynamics simulations. For example,
612   temperature of the system is defined by
613   \begin{equation}
614   \frac{3}{2} N k_B T = \langle \sum_{i=1}^N \frac{1}{2} m_i v_i \rangle,
# Line 605 | Line 621 | distributed randomly to the particles using a Maxwell-
621   The initial positions of the particles are chosen so that there is no
622   overlap of the particles. The initial velocities at first are
623   distributed randomly to the particles using a Maxwell-Boltzmann
624 < ditribution, and then shifted to make the total linear momentum of the
625 < system $0$.
624 > distribution, and then shifted to make the total linear momentum of
625 > the system $0$.
626  
627   The core of Molecular Dynamics simulations is the calculation of
628   forces and the integration algorithm. Calculation of the forces often
# Line 632 | Line 648 | is
648   equations of motion. The Taylor expansion of the position at time $t$
649   is
650   \begin{equation}
651 < q(t+\Delta t)= q(t) + v(t) \Delta t + \frac{f(t)}{2m}\Delta t^2 +
651 > q(t+\Delta t)= q(t) + \frac{p(t)}{m} \Delta t + \frac{f(t)}{2m}\Delta t^2 +
652          \frac{\Delta t^3}{3!}\frac{\partial^3 q(t)}{\partial t^3} +
653          \mathcal{O}(\Delta t^4)
654   \label{Ineq:verletFuture}
655   \end{equation}
656   for a later time $t+\Delta t$, and
657   \begin{equation}
658 < q(t-\Delta t)= q(t) - v(t) \Delta t + \frac{f(t)}{2m}\Delta t^2 -
658 > q(t-\Delta t)= q(t) - \frac{p(t)}{m} \Delta t + \frac{f(t)}{2m}\Delta t^2 -
659          \frac{\Delta t^3}{3!}\frac{\partial^3 q(t)}{\partial t^3} +
660          \mathcal{O}(\Delta t^4) ,
661   \label{Ineq:verletPrevious}
# Line 666 | Line 682 | code can be found in Ref.~\cite{Meineke2005}.
682   code can be found in Ref.~\cite{Meineke2005}.
683  
684   \subsection{Langevin Dynamics\label{In:ssec:ld}}
685 < In many cases, the properites of the solvent in a system, like the
686 < water in the lipid-water system studied in this dissertation, are less
687 < interesting to the researchers than the behavior of the
688 < solute. However, the major computational expense is ofter the
689 < solvent-solvent interation, this is due to the large number of the
690 < solvent molecules when compared to the number of solute molecules, the
691 < ratio of the number of lipid molecules to the number of water
692 < molecules is $1:25$ in our lipid-water system. The efficiency of the
693 < Molecular Dynamics simulations is greatly reduced, with up to 85\% of
694 < CPU time spent calculating only water-water interactions.
685 > In many cases, the properites of the solvent (like the water in the
686 > lipid-water system studied in this dissertation) are less interesting
687 > to the researchers than the behavior of the solute. However, the major
688 > computational expense is ofter the solvent-solvent interactions, this
689 > is due to the large number of the solvent molecules when compared to
690 > the number of solute molecules. The ratio of the number of lipid
691 > molecules to the number of water molecules is $1:25$ in our
692 > lipid-water system. The efficiency of the Molecular Dynamics
693 > simulations is greatly reduced, with up to 85\% of CPU time spent
694 > calculating only water-water interactions.
695  
696   As an extension of the Molecular Dynamics methodologies, Langevin
697   Dynamics seeks a way to avoid integrating the equations of motion for
# Line 728 | Line 744 | q_\alpha(0)}{\omega_\alpha} \sin(\omega_\alpha t),
744   q_\alpha(0)}{\omega_\alpha} \sin(\omega_\alpha t),
745   \label{Ineq:randomForceforGLE}
746   \end{equation}
747 < depends only on the initial conditions. The relationship of friction
748 < kernel $\xi(t)$ and random force $R(t)$ is given by the second
749 < fluctuation dissipation theorem,
747 > that depends only on the initial conditions. The relationship of
748 > friction kernel $\xi(t)$ and random force $R(t)$ is given by the
749 > second fluctuation dissipation theorem,
750   \begin{equation}
751   \xi(t) = \frac{1}{k_B T} \langle R(t)R(0) \rangle.
752   \label{Ineq:relationshipofXiandR}

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines