FENZI: GPU-enabled Molecular Dynamics Simulations of Large Membrane Regions based on the CHARMM force field and PME

Size: px
Start display at page:

Download "FENZI: GPU-enabled Molecular Dynamics Simulations of Large Membrane Regions based on the CHARMM force field and PME"

Transcription

1 211 IEEE International Parallel & Distributed Processing Symposium : GPU-enabled Molecular Dynamics Simulations of Large Membrane Regions based on the force field and PME Narayan Ganesan, Michela Taufer Dept. of Computer & Inf. Sciences University of Delaware {ganesan, taufer}@udel.edu Brad Bauer, Sandeep Patel Dept. of Chemistry and Biochemistry University of Delaware {babauer, patel}@udel.edu Abstract When studying membrane-bound protein receptors, it is necessary to move beyond the current state-of-theart simulations that only consider small membrane patches and implicit solvent. Limits of traditional computer platforms negatively impact the model s level of realism and the computational scales achievable. On the other hand, multi-core platforms such as GPUs offer the possibility to span length scales in membrane simulations much larger and with higher resolutions than before. To this end, this paper presents the design and implementation of an advanced GPU algorithm for Molecular Dynamics (MD) simulations of large membrane regions in the NVT, NVE, and NPT ensembles using explicit solvent and Particle Mesh Ewald (PME) method for treating the conditionallyconvergent electrostatic component of the classical force field. A key component of our algorithm is the redesign of the traditional PME method to better fit on the multithreading GPU architecture. This has been considered a fundamentally hard problem in the molecular dynamics community working on massively multithreaded architecture. Our algorithm is integrated in the code (yun dong de FEN ZI in Mandarin or moving molecules in English). The paper analyzes both the performance and accuracy of large-scale GPU-enabled simulations of membranes using, showing how our code can enable multi-nanosecond MD simulations per day, even when using PME. Keywords-Biomolecular Systems Simulation; Molecular Dynamics; DMPC Membranes; Performance; Accuracy; Ewald Summation; Particle Mesh Ewald. I. INTRODUCTION Roughly one-third of the human genome is composed of membrane-bound proteins that are only now becoming structurally resolved due to heroic experimental efforts. Furthermore, pharmaceuticals target membrane-bound protein receptors (such as G-protein coupled receptors, GPCR s), thus emphasizing the importance of such systems to human health and understanding of dysfunction. Computing and communication bottlenecks limit membrane simulations to the simulation of small regions (or patches) of physiological membranes using implicit solvent representations. Heterogeneity of membranes spans length scales much larger than included in these smaller model systems. Thus when studying membrane-bound protein receptors, it is necessary to move beyond the current state-of-the-art simulations that only consider small patches and implicit solvent. While the biological models are mature for this step, still limits of traditional computer platforms negatively impact the model s level of realism and the computational scales achievable. With the advent of multi-core platforms such as GPUs, scientists have now new computing tools to overcome these barriers. At the same time the migration of the biological models to the new platforms include new challenges, i.e., algorithm redesign and implementation to meet the new platform configurations. To this end, in this paper we present the design and implementation of an efficient algorithm for Molecular Dynamics (MD) simulations for GPUs. The algorithm is integrated in the code (yun dong de FEN ZI in Mandarin or moving molecules in English) that enables the fast simulation of membrane regions in the NVT (i.e., constant atom number, constant volume, and constant temperature) NVE (i.e., constant atom number, constant volume, and constant energy), and NPT (i.e., constant atom number, constant pressure, and constant temperature) ensembles with explicit solvent. A key contribution of this paper is the presentation of an efficient GPU algorithm for the Ewald summation for treating the conditionally-convergent electrostatic component of the classical force field and the Particle Mesh Ewald (PME) method to compute the corresponding reciprocal space terms. While adding accuracy to the simulation results, the use of the PME method also adds an additional layer of complexity to the efficient implementation of on GPUs, since requiring to deal with charge location and spread as well as Fast Fourier Transform (FFT) calculations on a multithreading platform. We address this challenge in by rethinking the way changes are spread and handled from a charge-centric representation (as it is traditionally done in CPU algorithms) to a lattice-centric representation. We use to analyze the performance and accuracy of large-scale, GPU-enabled computations of extended phospholipid bilayer membranes (DMPC), and show how our code can enable multi-nanosecond MD simulations per day, even when using PME for large membranes, with 273,936 atoms in explicit solvent. We use as the reference code to validate the accuracy and /11 $ IEEE DOI 1.119/IPDPS

2 performance by comparing results of versus results of the same DMPC simulations with on multicore systems. Even though the motivation of this work was the study of large membrane regions, is a full-fledged general purpose MD package based on canonical force fields and PME for GPU enabled simulation of a broad class of molecular systems. Thus, for the sake of comparison with other existing MD codes for GPU including PME, we measure performance when running the DHFR benchmark. The rest of the paper is organized as follows: Section II presents the modeling and implementation of our MD code, including the PME algorithm for GPUs; Section III shows the performance and accuracy of the code from the scientists and computer scientists prospectives; Section IV discusses relevant related work; and Section V concludes the paper and lists future work. A. MD Modeling II. MODELING AND IMPLEMENTATION (yun dong de FEN ZI in Mandarin or moving MOLECULES in English) enables GPU-based MD simulations in constant energy (also called NVE) or constant temperature (also called NVT) ensembles using a modified version of the force field in terms of force field functional forms and measurement units [1]. The entire MD simulations (i.e., intra-molecular and long range potentials) are computed on GPU. The intra-molecular potential, which includes bonds, angles including the Urey-Bradley summation, and dihedral, is as follows: V intra = K b (r r ) 2 + K a (θ θ ) 2 + bonds angles K UB (S S ) 2 + Urey-Bradley K φ (1 + cos(nφ δ)) + dihedrals K ω (ω ω ) 2 impropers where K b,k a,k UB,K φ,k ω are the bond, angle, Urey- Bradley, dihedral and improper force constants, and r,θ,s,δ,ω are their corresponding equilibrium distances. Non-bonded Lennard-Jones interactions are modeled using a standard 6-12 dispersion-repulsion potential: ( [ pairs (σij ) 12 ( ) ]) 6 σij V LJ = 4ɛ ij i,j where ɛ ij and σ ij are the potential well depth and the van der Waals radius, respectively, used in the Lennard-Jones potential. r ij r ij Long range electrostatic interactions can be computed either by using the reaction force field (RF) or the Ewald summation method. In this work we employ Ewald summation for electrostatic interactions and the PME method to compute the corresponding reciprocal space terms. In the Ewald summation method the electrostatic interactions are divided into the direct space energy (E dir ), the reciprocal space energy (E rec ), and the self energy (E corr )contributions to the total energy, depending on the distance of the interaction [2]. The three contributions are: E dir = N 1 i=1 E rec = 1 2πV E corr = 1 2 N j>i m (i,j) Excl q i q j erfc(βr ij ) r ij exp( π 2 m 2 /β 2 ) m 2 S( m )S( m ) q i q j erf(β r i r j ) r i r j β π i=1 Nq 2 i where β is the Ewald parameters and m =(m 1,m 2,m 3 ) are reciprocal space lattice vectors, V is the volume of the unit cell in the reciprocal space, and S( m ) is the lattice structure factor given by: S( m )= j q j exp( m r j ) The solvent is treated explicitly. Water bond, angle, and charge parameters are transferred directly from the SPC/Fw model of Wu et al. [3]. B. Code Implementation on GPU We use CUDA for our code implementation. Bond-, angle-, and dihedral interactions are each handled by a unified kernel that evaluates the potentials by iterating through the list of all atoms bonded to or involved in an angle with an atom i and accumulates the appropriate forces. Unlike non-bonded lists, the bond-, angle-, and dihedral lists never require updating, so they are constructed once on the CPU at the start of the simulation and then copied to the GPU. Non-bonded interactions (i.e., Lennard-Jones and direct space electrostatic terms) are handled by two kernels, NBBuild and NonBondForce. In NBBuild, usesthecellbasedapproachfor building the nonbond neighbor list. The entire domain of atoms is decomposed into regular cubic cells of edge length equal to r cut.thelistofalltheneighboringatomswithin the cutoff r cut is build so that only atoms in the neighboring cells are searched. The neighbor list is constructed using the Verlet list approach [4], in which a list is constructed for each atom containing all atoms within a cutoff r list >r cut and not in the exclusion. During the construction of the neighbor list, the exclusion list is rapidly searched by constructing abitmaskofexclusionatoms[5]foreachparticle,which enables bitwise operations. The nonbond neighbor list only

3 needs to be updated when an atom has moved beyond a user defined buffer size. While building the neighbor list for each atom, only the atoms within the current cell and 26 neighboring cells in 3-dimensional space need to be searched for atoms within the cutoff distance. The cells themselves are updated efficiently by carefully keeping track of movement atoms within the cells and employing atomic intrinsics where possible. In NonBondForce, each thread iterates through the list of neighbors for a single atom i and accumulates the interactions between i and all its neighbor list entries. The texture cache, which speeds up reading from global memory locations that are not contiguous, is used for reading the coordinates of the neighbor atoms. Switched force forms are used for the Lennard-Jones potential to ensure that both energies and forces go smoothly to zero at the cutoff r cut. C. Smooth Particle Mesh Ewald The multithreaded GPU architecture requires us to redesign the Smooth Particle Mesh Ewald (PME) algorithm, used to compute the corresponding reciprocal space terms in the Ewald summation. More in particular, we need to rethink the way charges are spread, handled, and updated during the MD simulation. This problem has been considered fundamentally hard to be ported onto massively multithreaded platforms [6] due to the atomic nature of computations involved. Here we propose to redesign the PME computation by moving away from a charge-centric representation (as it is traditionally done in CPU algorithms) to a lattice-centric representation for GPUs. This enables us to keep track of the movement of charges to different lattice points efficiently and optimally. The Smooth Particle Mesh Ewald (PME) component of the Ewald summation method (E rec ) requires us to deal with location of charges. The entire procedure involves: (1) spreading the charges within a neighboring volume of 4x4x4 cells for each charge in order to obtain a 3-dimensional charge matrix, (2) computing the inverse 3D FFT of the matrix, (3) multiplying the charge matrix by pre-computed structure constants, (4) computing the forward FFT of the product, and (5) summing the entries of the matrix to obtain energy and forces [2]. The charge spreading is performed in the ChargeSpread kernel and the list updating is done in the LatticeUpdate kernel. The FFT computations, i.e., inverse and forward FFTs, are performed using the CUDA FFT library. The kernel CUFFTExec performs both the FFTs. The multiplication of the charge matrix by precomputed structure constants is performed by the kernel BCMultiply. Last but not least, the summation of the entries of the matrix to obtain energy and forces is performed by the kernel PMEForce. Of these five steps, the charge spreading is one of the most compute intensive. In [7], where another example of MD code with PME for GPUs is presented, this step is Figure 1. Example of influence region of three charges. The combined effect of the multiple charges impacts the volume under intersection of neighborhoods. implemented by a three-step process which involves placing the charges on lattice points with utmost one charge per point (with extra charges considered separately), accumulating the impact of charges for each points, and finally accumulating the effects of the extra charges. Our implementation differs from the implementation in [7] since we accumulate impact of charges based on a local list of charges for each lattice point. Our charge spreading on GPU is performed by maintaining a local neighbor list for each lattice point, followed by an efficient update of list caused due to the movement of particles. In general, the charge contained in each atom is spread within a neighborhood of 4x4x4 lattice points around the particle. The spreading of charges is done with the help of the cardinal B-spline according to the summation: Q(k x,k y,k z )= N q i M 4 (u xi k x n 1 K x ) i=1 n 1,n 2,n 3 M 4 (u yi k y n 1 K y ) M 4 (u zi k z n 1 K z ) where k x,k y,k z index the 3-D charge matrix of size, K x,k y,k z and u xi,u yi,u zi are the relative x, y, z coordinates of the ith particle w.r.t. the charge matrix. The cardinal B-spline M 4 is continuously differentiable and has acompactsupport.asshowninfigure1,eachchargeis spread around a neighborhood of 4x4x4 lattice points and multiple charges can contribute to the total charge at each point. When a charge moves due to dynamics of the system, as shown in Figure 2 where a charge moves from Point 2 to Point 2, we can identify three regions of the lattice: (1) a region from which the charge is removed from the neighbor lists (light gray region in Figure 2), (2) a region whose neighbor lists gain the charge (dark grey region in Figure 2), and (3) a region whose neighbor lists are not impacted (white region in Figure 2). The solid arrow indicates the

4 Figure 2. Example of charge update for which only the affected lattice points are updated. Figure 3. When a single cell is affected due to displacement of multiple charges, it is updated with the help of integer atomic intrinsics. charge movement; dash arrows indicate mapping between lattice points. With reference to the mapping, the lattice points gaining the charge impact are initially not aware of the charge movement and thus need to search through the global list of charges for the information when an update takes place. This search is time demanding. At the same time, we observe that the movement of a charge produces an equal number of lists gaining the charge and lists losing the charge. The latter lists are associated to lattice points that are aware of the destination points in a one-to-one mapping schema showed in Figure 2. We can benefit from this one-toone mapping to reduce the time for the update of the lists by having the threads of the points losing the charge updating the lists of the points gaining the charge. This will prevent the lattice points gaining the charge impact to explore the whole global list of charges. In case multiple charges from different lattice points move into the neighborhood of the same lattice point, then the neighbor list of that point is updated by multiple threads. In Figure 3 the local neighbor list of a point point is updated by two different threads due to the movement of two different charges (Charges 2 and 1). Since this update requires the addition of two charges by different threads to the same neighbor list, this is performed with the help of thread safe integer atomic intrinsics. To minimize the number of floating point calculations, we pre-compute several functions (i.e., B-Spline values, the structure constants, and the value of error function(erfc)) and store them in the constant memory. III. PERFORMANCE AND ACCURACY A. Benchmark and Set-up To demonstrate both scalability and performance across multiple scales of system sizes, we considered a commonly used membrane system, the lipid bilayer membrane (DMPC), with three different sizes, one four times larger than the previous, as shown in Figure 4: Small membrane (DMPC 1x1): 46.8 X 46.8 X 76. Å, 17,4 atoms (i.e., 14,96 bonds, 19,18 angles, and 22,536 dihedrals) Medium membrane (DMPC 2x2): 93.6 X 93.6 X 76. Å, 68,484 atoms (i.e., 56,696 bonds, 76,588 angles, and 9,144 dihedrals) Large membrane (DMPC (4x4): 187.2Å X 187.2Å X 76.Å, 273,936 atoms (i.e., 226,784 bonds, 36,352 angles, 36,576 dihedrals) In the small membrane (DMPC 1x1) we deal with 2,836 explicit water molecules; in the medium membrane we deal with 11,5 explicit water molecules; in the large membrane we deal with 46,863 explicit water molecules. Normally Lipid bilayer membranes are an important class of biological components, and fundamental study of their structure, dynamics, and interactions with peptides, proteins, and medicinally-relevant small molecules is important. Current state-of-the-art simulations employ traditional supercomputers and study regions of the size of the small DMPC, because of time and resource constrains. We ran the same DMPC simulations (with the same input files - i.e., coordinates, topology, and input parameters) on aclusternodeusing[1]asareferenceforassessing both accuracy and performance. simulations were run on 1, 2, 4, and 8 cores of an Intel Xeon with speed 2.6GHz and 8GB of memory. We did not scale beyond 8 cores of a single node because the cluster used was not provided with InfiniBand and thus the performance was very poor across nodes. For our GPU simulations, we used both a GTX 48 GPU (Fermi) with 48 cores and 1.5 GB memory and a C25 GPU (Fermi) with 448 cores and 3 GB memory. The performance values presented are the average values computed over three repeated simulations. The standard deviation is not reported because it is close to zero

5 Figure 4. The three DMPC membranes considered in this paper - each membrane is four times larger than the previous membrane. B. Performance When addressing performance, scientists want to know how much simulated time (e.g., in nanoseconds or microseconds) can be performed in one day. Flops and MIPS have little meaning for the scientists. Thus, in this paper we measure performance and compare it against other codes in terms of nanoseconds per day. First of all, we compared performance versus performance, since is the code we normally use for our membrane studies. For our comparison, we ran 1, steps of a constant energy MD simulation (NVE) with the smaller membrane (DMPC 1x1) using a r cutoff of 8Å withabuffercutoffr list for the list updates of 9.5Å. Each step was 1 fs long. For the GPU platform, we considered both a GTX 48 and a C25. Figure 5 shows the number of nanoseconds per day for the double-precision parallel code optimized for 1, 2, 4, and 8 CPU cores versus the number of nanoseconds per day of on one single GPU. uses MPI for distributing the membrane domain across cores and was optimized for the specific node. The figure shows how the speedup of one single-precision GPU is up to 1X faster than a 8-core, double-precision CPU node. To study the scalability of our code, we compared its performance in terms of ns/day for the three lipid bilayer membranes (DMPC) systems with different sizes. Each DMPC membrane is four times larger than the previous. simulations were performed on the GTX 48 and C25 GPU. Figure 6 shows how enables multinanosecond MD simulations per day with the considered membranes. More in particular, for the small membrane and the faster GTX 48 GPU, our MD code running on GPUs achieves simulation rates of up to million MD steps with 3.78 ms per step (22.86 nanoseconds per day with a MD step size of 1 fs), up to 6.79 million MD steps per day nsec/day GTX 48 C25 (#cores on cluster node) 17.7 (single GPU) Figure 5. Comparison of performance in terms of ns/day for on 1, 2, 4, and 8 CPU cores versus on GTX 48 (Fermi chip). with 12.7 ms per step (6.79 nanoseconds per day with a MD step size of 1 fs) for the medium membrane, and up to 1.63 million MD steps per day with ms per step (1.63 nanoseconds per day with a MD step size of 1 fs) for the large membrane. The performance on C25 GPUs is slightly slower but still very competitive compared with traditional and more expensive supercomputers. Overall, allows us to simulate larger membranes, larger than simple regions, over a longer simulated time interval in a significantly shorter turnaround time. Work in progress includes the study of the membrane properties - i.e., structural (densities, electron density profiles), order pa

6 ns/day DMPC 1x1 DMPC 2x2 DMPC 4x4 GTX 48 C25 energy fluctuations (kcal/mol) time step size (fs) Figure 6. Simulations of three lipid bilayer membranes (DMPC) with three different sizes, each four times larger than the previous. Figure 7. Plot of the total energy fluctuations of DMPC 1x1 as a function of time step size for different time step size - i.e.,.1 fs,.2 fs,.5fs,1 fs, and 2 fs and single precision. rameters (SCD) and electrostatic properties (dipole potential, water dipole moments), as well as orientational properties of water - over a time scale of tens of hundreds nanoseconds, to assess whether properties observed in small membranes with simulations on traditional CPU clusters are still true as the membranes significantly grow in size. C. Accuracy Clearly, even the fastest MD code has little value if the results it produces are not accurate. To ensure that our simulation results are accurate, we first ran the same 1 ps MD simulation for the smaller membrane system in equilibrium with different time step size - i.e.,.1 fs,.2 fs,.5 fs, 1 fs, and 2 fs - and single precision. We plot the fluctuations in the total energy as a function of the time step size in Figure 7. According to Allen and Tildesley [4], a plot of the energy fluctuations versus time step size should follow an approximately logarithmic trend. In Figure 7 we observe that, for large time step size (larger than.5 fs), the fluctuations in total energy of MD simulations are proportional to the time step size. On the other hand, we observe a different behavior for step size less than.5 fs. This is consistent with results previously presented and discussed in [8]. Next, we ran simulations with the smaller membrane (DMPC 1x1) system in equilibrium on CPU using the double-precision code and on GPU using our single precision code. The simulations are performed in constant temperature. As pointed out in [9], energy drift can be observed in long constant energy simulations and the drift can be easily overcome by maintaining constant temperature (e.g., using velocity reassignment which effectively changes the energy state of the system at specified intervals, thereby eliminating the drift for production simulations) Figure 8 plots the temperature (a function of kinetic energy), bond energies (i.e., bond-, angle, Urey-Bradley-, improper-, and dihedral energies), and non-bond energies (i.e., VDW, electrostatic, and PME) over a 3 ns simulation time for the small membrane. The several quantities (i.e., temperature and energy) fluctuate, as expected, around the same average value for both the double precision CPU simulation and single precision GPU simulation. We observed that initial energy (step ) matches the initial energy. During the initial phase of the simulation, before the system reaches the equilibrium, we also observed some energy drifting due to the different thermostats used by the two codes (i.e., uses Langevin thermostat while uses velocity reassignment). Eventually the two simulations converge, indicating that does in fact produce results consistent with and the use of double precision is not needed on GPUs for this type of simulations. A detailed analysis of the energy profiles from the science point of view is work in progress. IV. RELATED WORK Other scientists have been targeting the design and implementation of MD codes including the PME computation entirely on GPUs. One of the most relevant published work is ACEMD [7]. As outlined in Section II-C, is different from ACEMD in terms of how it deals with the charges and charge spreading for PME. Since a profiling of

7 x Temperature Bond Angle Urey Bradley Improper Dihed x VDW 4 3 Electrostatic PME Figure 8. Comparison of the temperature and energy profiles for 3 ns of NVT MD simulation with 1fs step size for on single core, 64 bits and on GTX 48, 32 bits. the individual kernels for ACEMD is not feasible (the code is not open-source) and this information is not available (data in [7] is not readable for comparison), we cannot quantify the impact of our kernel performing the charge spreading versus the equivalent ACEMD kernel. As also outlined in [1], ACEMD has been designed and optimized for single GPUs and small molecular systems. ACEMD has been parallelized by distributing the kernels across three GPUs in a task-based parallelism. This approach can limit the size of a molecular system that can be ultimately simulated with ACEMD to approximately 1K atoms, due to global memory limitations. On the contrary, while designing as described in Section II, we have been using data structures and algorithms that can be easily parallelized in a data parallelism way. We are currently working on a parallel version of that uses domain decomposition to assure scalability of the simulation for very large molecular systems supported by multiple GPUs. Walker and co-workers are also currently working on a version of AMBER for GPUs including PME. Preliminary performance results for the DHFR (23K atoms) and ApoA1 (92K atoms) benchmarks are presented in their webpage [11] but a manuscript reporting their algorithm is still work in progress. GROMACS [6] also supports Particle-Mesh-Ewald (PME). As, GROMACS uses the mathematical models in [2] for the implementation of the PME method, however we were not able to find a manuscript presenting the algorithmic implementation of its PME code on GPU. To compare performance with other codes performing PME entirely on GPU (i.e., ACEMD and AMBER), we ran the MD simulation of a well-known benchmark such as the DHFR system (23K atoms with 16,569 bonds, 11,584 angles, and 6,71 dihedrals) using initial parameters that are similar, when possible, to the parameters used by the other codes in their webpages, i.e., 8 Ånonbondcutoff, a buffer cutoff of 9.5 Å, 1, MD steps in an NVE ensemble, and with both 1 fs and 2 fs time step. Figure 9 shows the performance of DHFR MD simulations with on GTX 48 and C25 (Fermi chip). Performance results on ACEMD and AMBER are presented in [12] and in [11] respectively. While comparing these results with our results, we observe that gives better performance than AMBER (i.e., according to [11], AMBER reaches ns/day on a C25 with a MD step of 2 fs) and

8 GTX 48 C25 2fs 1fs 2fs 1fs MD simulation of DHFR on C25 and GTX 48 (Fermi) ns/day Figure 9. performance in ns/day of DHFR simulations using different time steps and GPUs. similar performance as ACEMD (i.e., from our testing of ACEMD on GTX 48, the code reaches ns/day with amdstepof2fs).notethatourmdalgorithmisdifferent than the ACEMD algorithm. As outlined above, we deal with the PME charge spreading in a significantly different way that ACEMD. Moreover, the nonbond interactions in still use exclusion and non-bond lists as in codes such as, while ACEMD does no longer refer to these data structures in [7]. GROMACS performance in [13] indicates values below 2 ns/day for the DHFR system on a C25 GPU, however the webpage does not include important information such as cutoff values and other parameter settings. Thus we cannot fairly compare these values with performance. Coarse-grained MD simulations of large membranes for long time intervals (of the order of 1 ns) are also presented in [14]. The coarse-grained nature of these simulations results in significant loss of information and a lower resolution than our code. Up to a few microseconds of MD simulations of small membrane can be reached on special purpose architectures such as Anton [15]. Such performance is feasible on Anton because of multi-step integration, less frequent long distance force calculations, fixed precision calculations, and small FFT grid size (32x32x32). At the same time these factors lead to loss of resolution and accuracy. Moreover, the size of the hardware defines the size of molecular systems being simulated. The size of the molecular system has a significant impact on Anton s performance: the simulation of larger systems can lead to a loss in performance. The application specific nature of the architecture does not allow for flexibility in terms of FFT grid size, choice of crystal lattice (cubic dimensions), and spline functions. This level of flexibility is possible on general purpose GPUs. Last but not least, we do not compare with the GPU-based NAMD [16] since the current version of NAMD for GPUs does not perform PME on GPUs, thus using different algorithm for the electrostatic interactions. V. CONCLUSION In this paper we presented the design and implementation of, a MD code for the simulation of large membrane regions in both NVT and NVE ensembles. A key contribution of the paper is the PME algorithm that we redesigned from a charge-centric to a lattice-of-points-centric prospective to better fit the multithreading GPU architecture. enables multi-nanosecond MD simulations per day, even when using PME for large membranes. achieves simulation rates of up to 22.8 nanoseconds per day with a MD step size of 1 fs for a small DMPC membrane of 17,4 atoms, up to 6.7 nanoseconds per day with a MD step size of 1 fs for a medium DMPC membrane of 68,484 atoms, and up to 1.6 nanoseconds per day with a MD step size of 1 fs for a large DMPC membrane of 273,936 atoms; all membranes were in explicit solvent. Because of our general design and implementation approach, is full-fledged general purpose MD package based on canonical force fields and PME for GPU enabled simulation of a broad class of molecular systems. Work in progress includes the analysis of properties such as structural (densities, electron density profiles), order parameters (SCD) and electrostatic properties (dipole potential, water dipole moments), as well as orientational properties of water over a time scale of tens of hundreds nanoseconds. Our final goal is to assess whether properties observed in small membranes with simulations on traditional CPU clusters are still true as the membranes significantly grow in size. By speeding up the MD simulation of large membranes, allows us to answer this question accurately in a short turnaround time. ACKNOWLEDGMENT This work was supported by the National Science Foundation, grant # and grant #922657, by the U.S. Army, grant #ARO CS, and by the NVIDIA University Professor Partnership Program. The authors want to thank Peng Wang (NVIDIA) and Shucai Xiao (UVA) for their feedback on optimization techniques in CUDA. REFERENCES [1] B. Brooks, et al.,: A Program for Macromolecular Energy, Minimization, and Dynamics Calculations. J. Comp. Chem., 4, , [2] U. Essman, et al., A Smooth Particle Mesh Ewald Potential. J. Chem. Phys.. 13, ,

9 [3] Y. Wu, H. L. Tepper, and G. Voth, A Flexible Simple Point Charge Water Model with Improved Liquid State Properties. J. Chem. Phys., 124, 2453, 26. [4] M. P. Allen and D. J. Tildesley, Computer Simulation of Liquids. Oxford: Clarendon Press, [5] J. E. Stone, et al., Accelerating Molecular Modeling Applications with Graphics Processors. J of Comp. Chemistry, 28(16), , 27. [6] D. van der Spoel, et al., GROMACS: Fast, Flexible, and Free. J. Comput. Chem. 26, , 25. [7] M. Harvey, G. Giupponi and G. De Fabritiis, ACEMD: Accelerating Biomolecular Dynamics in the Microsecond Time Scale. J. Chem. Theory and Comput. 5, , 29. [8] B. A. Bauer, J. E. Davis, M. Taufer, and S. Patel, Molecular Dynamics Simulations of Aqueous Ions at the Liquid-Vapor Interface Accelerated Using Graphics Processors, J.Comp. Chem., 32(3), , 211. [9] M. S. Friedrichs, et al., Accelerating Molecular Dynamic Simulation on Graphics Processor Units, J. Comput. Chem., 3, , 29. [1] J. E. Stone, D. J. Hardya, I. S. Ufimtsevb, K. Schulten GPUaccelerated Molecular Modeling Coming of Age. J. Molecular Graphics and Modeling, 29, , 21. [11] S. Le Grand, A. W. Goetz, D. Xu, D. Poole and R. C. Walker, URL: [12] M. Harvey, G. Giupponi and G. De Fabritiis, URL: [13] GROMACS Performance on GPU, URL: [14] B. A. Hall and M. S. P. Sansom, Coarse-Grained MD Simulations and Protein-Protein Interactions: The Cohesin- Dockerin System. J. Chem. Theory Comput., 5(9), , 29. [15] D E. Shaw, et al., Millisecond-scale Molecular Dynamics Simulations on Anton. In Proceedings of the ACM/IEEE Conference on Supercomputing (SC9), November 14-2, 29. [16] J. C. Phillips, et al., Scalable Molecular Dynamics with NAMD. J. of Comp. Chemistry, 26, ,

Efficient Parallelization of Molecular Dynamics Simulations on Hybrid CPU/GPU Supercoputers

Efficient Parallelization of Molecular Dynamics Simulations on Hybrid CPU/GPU Supercoputers Efficient Parallelization of Molecular Dynamics Simulations on Hybrid CPU/GPU Supercoputers Jaewoon Jung (RIKEN, RIKEN AICS) Yuji Sugita (RIKEN, RIKEN AICS, RIKEN QBiC, RIKEN ithes) Molecular Dynamics

More information

Molecular dynamics simulation. CS/CME/BioE/Biophys/BMI 279 Oct. 5 and 10, 2017 Ron Dror

Molecular dynamics simulation. CS/CME/BioE/Biophys/BMI 279 Oct. 5 and 10, 2017 Ron Dror Molecular dynamics simulation CS/CME/BioE/Biophys/BMI 279 Oct. 5 and 10, 2017 Ron Dror 1 Outline Molecular dynamics (MD): The basic idea Equations of motion Key properties of MD simulations Sample applications

More information

Parallelization of Molecular Dynamics (with focus on Gromacs) SeSE 2014 p.1/29

Parallelization of Molecular Dynamics (with focus on Gromacs) SeSE 2014 p.1/29 Parallelization of Molecular Dynamics (with focus on Gromacs) SeSE 2014 p.1/29 Outline A few words on MD applications and the GROMACS package The main work in an MD simulation Parallelization Stream computing

More information

Advanced Molecular Molecular Dynamics

Advanced Molecular Molecular Dynamics Advanced Molecular Molecular Dynamics Technical details May 12, 2014 Integration of harmonic oscillator r m period = 2 k k and the temperature T determine the sampling of x (here T is related with v 0

More information

The Fast Multipole Method in molecular dynamics

The Fast Multipole Method in molecular dynamics The Fast Multipole Method in molecular dynamics Berk Hess KTH Royal Institute of Technology, Stockholm, Sweden ADAC6 workshop Zurich, 20-06-2018 Slide BioExcel Slide Molecular Dynamics of biomolecules

More information

The Molecular Dynamics Method

The Molecular Dynamics Method The Molecular Dynamics Method Thermal motion of a lipid bilayer Water permeation through channels Selective sugar transport Potential Energy (hyper)surface What is Force? Energy U(x) F = d dx U(x) Conformation

More information

Optimizing GROMACS for parallel performance

Optimizing GROMACS for parallel performance Optimizing GROMACS for parallel performance Outline 1. Why optimize? Performance status quo 2. GROMACS as a black box. (PME) 3. How does GROMACS spend its time? (MPE) 4. What you can do What I want to

More information

An FPGA Implementation of Reciprocal Sums for SPME

An FPGA Implementation of Reciprocal Sums for SPME An FPGA Implementation of Reciprocal Sums for SPME Sam Lee and Paul Chow Edward S. Rogers Sr. Department of Electrical and Computer Engineering University of Toronto Objectives Accelerate part of Molecular

More information

LAMMPS Performance Benchmark on VSC-1 and VSC-2

LAMMPS Performance Benchmark on VSC-1 and VSC-2 LAMMPS Performance Benchmark on VSC-1 and VSC-2 Daniel Tunega and Roland Šolc Institute of Soil Research, University of Natural Resources and Life Sciences VSC meeting, Neusiedl am See, February 27-28,

More information

Introduction to molecular dynamics

Introduction to molecular dynamics 1 Introduction to molecular dynamics Yves Lansac Université François Rabelais, Tours, France Visiting MSE, GIST for the summer Molecular Simulation 2 Molecular simulation is a computational experiment.

More information

Efficient Molecular Dynamics on Heterogeneous Architectures in GROMACS

Efficient Molecular Dynamics on Heterogeneous Architectures in GROMACS Efficient Molecular Dynamics on Heterogeneous Architectures in GROMACS Berk Hess, Szilárd Páll KTH Royal Institute of Technology GTC 2012 GROMACS: fast, scalable, free Classical molecular dynamics package

More information

Molecular Dynamics Simulations. Dr. Noelia Faginas Lago Dipartimento di Chimica,Biologia e Biotecnologie Università di Perugia

Molecular Dynamics Simulations. Dr. Noelia Faginas Lago Dipartimento di Chimica,Biologia e Biotecnologie Università di Perugia Molecular Dynamics Simulations Dr. Noelia Faginas Lago Dipartimento di Chimica,Biologia e Biotecnologie Università di Perugia 1 An Introduction to Molecular Dynamics Simulations Macroscopic properties

More information

Computer simulation methods (2) Dr. Vania Calandrini

Computer simulation methods (2) Dr. Vania Calandrini Computer simulation methods (2) Dr. Vania Calandrini in the previous lecture: time average versus ensemble average MC versus MD simulations equipartition theorem (=> computing T) virial theorem (=> computing

More information

Why Proteins Fold? (Parts of this presentation are based on work of Ashok Kolaskar) CS490B: Introduction to Bioinformatics Mar.

Why Proteins Fold? (Parts of this presentation are based on work of Ashok Kolaskar) CS490B: Introduction to Bioinformatics Mar. Why Proteins Fold? (Parts of this presentation are based on work of Ashok Kolaskar) CS490B: Introduction to Bioinformatics Mar. 25, 2002 Molecular Dynamics: Introduction At physiological conditions, the

More information

What is Classical Molecular Dynamics?

What is Classical Molecular Dynamics? What is Classical Molecular Dynamics? Simulation of explicit particles (atoms, ions,... ) Particles interact via relatively simple analytical potential functions Newton s equations of motion are integrated

More information

The Computational Microscope

The Computational Microscope The Computational Microscope Computational microscope views at atomic resolution... Rs SER RER C E M RER N GA L... how living cells maintain health and battle disease John Stone Our Microscope is Made

More information

All-atom Molecular Mechanics. Trent E. Balius AMS 535 / CHE /27/2010

All-atom Molecular Mechanics. Trent E. Balius AMS 535 / CHE /27/2010 All-atom Molecular Mechanics Trent E. Balius AMS 535 / CHE 535 09/27/2010 Outline Molecular models Molecular mechanics Force Fields Potential energy function functional form parameters and parameterization

More information

The Molecular Dynamics Simulation Process

The Molecular Dynamics Simulation Process The Molecular Dynamics Simulation Process For textbooks see: M.P. Allen and D.J. Tildesley. Computer Simulation of Liquids.Oxford University Press, New York, 1987. D. Frenkel and B. Smit. Understanding

More information

GPU Acceleration of Cutoff Pair Potentials for Molecular Modeling Applications

GPU Acceleration of Cutoff Pair Potentials for Molecular Modeling Applications GPU Acceleration of Cutoff Pair Potentials for Molecular Modeling Applications Christopher Rodrigues, David J. Hardy, John E. Stone, Klaus Schulten, Wen-Mei W. Hwu University of Illinois at Urbana-Champaign

More information

JASS Modeling and visualization of molecular dynamic processes

JASS Modeling and visualization of molecular dynamic processes JASS 2009 Konstantin Shefov Modeling and visualization of molecular dynamic processes St Petersburg State University, Physics faculty, Department of Computational Physics Supervisor PhD Stepanova Margarita

More information

Bioengineering 215. An Introduction to Molecular Dynamics for Biomolecules

Bioengineering 215. An Introduction to Molecular Dynamics for Biomolecules Bioengineering 215 An Introduction to Molecular Dynamics for Biomolecules David Parker May 18, 2007 ntroduction A principal tool to study biological molecules is molecular dynamics simulations (MD). MD

More information

Development of Molecular Dynamics Simulation System for Large-Scale Supra-Biomolecules, PABIOS (PArallel BIOmolecular Simulator)

Development of Molecular Dynamics Simulation System for Large-Scale Supra-Biomolecules, PABIOS (PArallel BIOmolecular Simulator) Development of Molecular Dynamics Simulation System for Large-Scale Supra-Biomolecules, PABIOS (PArallel BIOmolecular Simulator) Group Representative Hisashi Ishida Japan Atomic Energy Research Institute

More information

Non-bonded interactions

Non-bonded interactions speeding up the number-crunching Marcus Elstner and Tomáš Kubař May 8, 2015 why care? key to understand biomolecular structure and function binding of a ligand efficiency of a reaction color of a chromophore

More information

Molecular Dynamics Simulation of a Nanoconfined Water Film

Molecular Dynamics Simulation of a Nanoconfined Water Film Molecular Dynamics Simulation of a Nanoconfined Water Film Kyle Lindquist, Shu-Han Chao May 7, 2013 1 Introduction The behavior of water confined in nano-scale environment is of interest in many applications.

More information

Folding WT villin in silico Three folding simulations reach native state within 5-8 µs

Folding WT villin in silico Three folding simulations reach native state within 5-8 µs Modeling of Cryo-EM Maps Workshop Baylor College of Medicine Klaus Schulten, U. Illinois at Urbana-Champaign Molecular Modeling Flexible Fitting 1: Introduction to Molecular Dynamics MD force field Equation

More information

Multiscale simulations of complex fluid rheology

Multiscale simulations of complex fluid rheology Multiscale simulations of complex fluid rheology Michael P. Howard, Athanassios Z. Panagiotopoulos Department of Chemical and Biological Engineering, Princeton University Arash Nikoubashman Institute of

More information

Non-bonded interactions

Non-bonded interactions speeding up the number-crunching continued Marcus Elstner and Tomáš Kubař December 3, 2013 why care? number of individual pair-wise interactions bonded interactions proportional to N: O(N) non-bonded interactions

More information

A microsecond a day keeps the doctor away: Efficient GPU Molecular Dynamics with GROMACS

A microsecond a day keeps the doctor away: Efficient GPU Molecular Dynamics with GROMACS GTC 20130319 A microsecond a day keeps the doctor away: Efficient GPU Molecular Dynamics with GROMACS Erik Lindahl erik.lindahl@scilifelab.se Molecular Dynamics Understand biology We re comfortably on

More information

Modeling and visualization of molecular dynamic processes

Modeling and visualization of molecular dynamic processes JASS 2009 Konstantin Shefov Modeling and visualization of molecular dynamic processes St-Petersburg State University Physics faculty Department of Computational Physics Supervisor PhD Stepanova Margarita

More information

Accelerating Protein Coordinate Conversion using GPUs

Accelerating Protein Coordinate Conversion using GPUs Accelerating Protein Coordinate Conversion using GPUs Mahsa Bayati Jaydeep P. Bardhan Department of Electrical and Department of Electrical and Computer Engineering Computer Engineering Northeastern University

More information

Multiple time step Monte Carlo

Multiple time step Monte Carlo JOURNAL OF CHEMICAL PHYSICS VOLUME 117, NUMBER 18 8 NOVEMBER 2002 Multiple time step Monte Carlo Balázs Hetényi a) Department of Chemistry, Princeton University, Princeton, NJ 08544 and Department of Chemistry

More information

Applications of the Spring. Force: molecules

Applications of the Spring. Force: molecules Applications of the Spring Atoms interact via the coulomb force Force: molecules When atoms are far apart, they are attractive When atoms are too close, they are repulsive Atoms in a molecule display relative

More information

A Nobel Prize for Molecular Dynamics and QM/MM What is Classical Molecular Dynamics? Simulation of explicit particles (atoms, ions,... ) Particles interact via relatively simple analytical potential

More information

arxiv: v1 [hep-lat] 7 Oct 2010

arxiv: v1 [hep-lat] 7 Oct 2010 arxiv:.486v [hep-lat] 7 Oct 2 Nuno Cardoso CFTP, Instituto Superior Técnico E-mail: nunocardoso@cftp.ist.utl.pt Pedro Bicudo CFTP, Instituto Superior Técnico E-mail: bicudo@ist.utl.pt We discuss the CUDA

More information

Principles and Applications of Molecular Dynamics Simulations with NAMD

Principles and Applications of Molecular Dynamics Simulations with NAMD Principles and Applications of Molecular Dynamics Simulations with NAMD Nov. 14, 2016 Computational Microscope NCSA supercomputer JC Gumbart Assistant Professor of Physics Georgia Institute of Technology

More information

arxiv: v1 [physics.chem-ph] 8 Mar 2010

arxiv: v1 [physics.chem-ph] 8 Mar 2010 arxiv:1003.1678v1 [physics.chem-ph] 8 Mar 2010 A new battery-charging method suggested by molecular dynamics simulations Ibrahim Abou Hamad 1,, M. A. Novotny 1,2, D. Wipf 3, and P. A. Rikvold 4 1 HPC 2,

More information

Accelerating Rosetta with OpenMM. Peter Eastman. RosettaCon, August 5, 2010

Accelerating Rosetta with OpenMM. Peter Eastman. RosettaCon, August 5, 2010 Accelerating Rosetta with OpenMM Peter Eastman RosettaCon, August 5, 2010 What is OpenMM? OpenMM is a library for molecular modeling on high performance architectures. Performance 600 500 502 400 ns/day

More information

XXL-BIOMD. Large Scale Biomolecular Dynamics Simulations. onsdag, 2009 maj 13

XXL-BIOMD. Large Scale Biomolecular Dynamics Simulations. onsdag, 2009 maj 13 XXL-BIOMD Large Scale Biomolecular Dynamics Simulations David van der Spoel, PI Aatto Laaksonen Peter Coveney Siewert-Jan Marrink Mikael Peräkylä Uppsala, Sweden Stockholm, Sweden London, UK Groningen,

More information

RWTH Aachen University

RWTH Aachen University IPCC @ RWTH Aachen University Optimization of multibody and long-range solvers in LAMMPS Rodrigo Canales William McDoniel Markus Höhnerbach Ahmed E. Ismail Paolo Bientinesi IPCC Showcase November 2016

More information

Molecular dynamics simulation of Aquaporin-1. 4 nm

Molecular dynamics simulation of Aquaporin-1. 4 nm Molecular dynamics simulation of Aquaporin-1 4 nm Molecular Dynamics Simulations Schrödinger equation i~@ t (r, R) =H (r, R) Born-Oppenheimer approximation H e e(r; R) =E e (R) e(r; R) Nucleic motion described

More information

Molecular Dynamics. A very brief introduction

Molecular Dynamics. A very brief introduction Molecular Dynamics A very brief introduction Sander Pronk Dept. of Theoretical Physics KTH Royal Institute of Technology & Science For Life Laboratory Stockholm, Sweden Why computer simulations? Two primary

More information

Supporting information

Supporting information Electronic Supplementary Material ESI for Physical Chemistry Chemical Physics. This journal is the Owner Societies 2018 Supporting information Understanding three-body contributions to coarse-grained force

More information

COMPARISON OF CPU AND GPU IMPLEMENTATIONS OF THE LATTICE BOLTZMANN METHOD

COMPARISON OF CPU AND GPU IMPLEMENTATIONS OF THE LATTICE BOLTZMANN METHOD XVIII International Conference on Water Resources CMWR 2010 J. Carrera (Ed) c CIMNE, Barcelona, 2010 COMPARISON OF CPU AND GPU IMPLEMENTATIONS OF THE LATTICE BOLTZMANN METHOD James.E. McClure, Jan F. Prins

More information

Parallel Utility for Modeling of Molecular Aggregation

Parallel Utility for Modeling of Molecular Aggregation Parallel Utility for Modeling of Molecular Aggregation Peter Spijker, Bart Markvoort and Peter Hilbers BioModeling and bioinformatics Eindhoven University of Technology Eindhoven,The Netherlands p.spijker@pumma.nl

More information

ab initio Electronic Structure Calculations

ab initio Electronic Structure Calculations ab initio Electronic Structure Calculations New scalability frontiers using the BG/L Supercomputer C. Bekas, A. Curioni and W. Andreoni IBM, Zurich Research Laboratory Rueschlikon 8803, Switzerland ab

More information

GPU acceleration of Newton s method for large systems of polynomial equations in double double and quad double arithmetic

GPU acceleration of Newton s method for large systems of polynomial equations in double double and quad double arithmetic GPU acceleration of Newton s method for large systems of polynomial equations in double double and quad double arithmetic Jan Verschelde joint work with Xiangcheng Yu University of Illinois at Chicago

More information

Potential Energy (hyper)surface

Potential Energy (hyper)surface The Molecular Dynamics Method Thermal motion of a lipid bilayer Water permeation through channels Selective sugar transport Potential Energy (hyper)surface What is Force? Energy U(x) F = " d dx U(x) Conformation

More information

Scalable and Power-Efficient Data Mining Kernels

Scalable and Power-Efficient Data Mining Kernels Scalable and Power-Efficient Data Mining Kernels Alok Choudhary, John G. Searle Professor Dept. of Electrical Engineering and Computer Science and Professor, Kellogg School of Management Director of the

More information

Lecture 11: Potential Energy Functions

Lecture 11: Potential Energy Functions Lecture 11: Potential Energy Functions Dr. Ronald M. Levy ronlevy@temple.edu Originally contributed by Lauren Wickstrom (2011) Microscopic/Macroscopic Connection The connection between microscopic interactions

More information

Introduction to numerical computations on the GPU

Introduction to numerical computations on the GPU Introduction to numerical computations on the GPU Lucian Covaci http://lucian.covaci.org/cuda.pdf Tuesday 1 November 11 1 2 Outline: NVIDIA Tesla and Geforce video cards: architecture CUDA - C: programming

More information

Welcome to MCS 572. content and organization expectations of the course. definition and classification

Welcome to MCS 572. content and organization expectations of the course. definition and classification Welcome to MCS 572 1 About the Course content and organization expectations of the course 2 Supercomputing definition and classification 3 Measuring Performance speedup and efficiency Amdahl s Law Gustafson

More information

Hands-on : Model Potential Molecular Dynamics

Hands-on : Model Potential Molecular Dynamics Hands-on : Model Potential Molecular Dynamics OUTLINE 0. DL_POLY code introduction 0.a Input files 1. THF solvent molecule 1.a Geometry optimization 1.b NVE/NVT dynamics 2. Liquid THF 2.a Equilibration

More information

The Molecular Dynamics Method

The Molecular Dynamics Method H-bond energy (kcal/mol) - 4.0 The Molecular Dynamics Method Fibronectin III_1, a mechanical protein that glues cells together in wound healing and in preventing tumor metastasis 0 ATPase, a molecular

More information

Scientific Computing II

Scientific Computing II Scientific Computing II Molecular Dynamics Simulation Michael Bader SCCS Summer Term 2015 Molecular Dynamics Simulation, Summer Term 2015 1 Continuum Mechanics for Fluid Mechanics? Molecular Dynamics the

More information

Graphics Card Computing for Materials Modelling

Graphics Card Computing for Materials Modelling Graphics Card Computing for Materials Modelling Case study: Analytic Bond Order Potentials B. Seiser, T. Hammerschmidt, R. Drautz, D. Pettifor Funded by EPSRC within the collaborative multi-scale project

More information

Theoretical comparative study on hydrogen storage of BC 3 and carbon nanotubes

Theoretical comparative study on hydrogen storage of BC 3 and carbon nanotubes J. At. Mol. Sci. doi: 10.4208/jams.121011.011412a Vol. 3, No. 4, pp. 367-374 November 2012 Theoretical comparative study on hydrogen storage of BC 3 and carbon nanotubes Xiu-Ying Liu a,, Li-Ying Zhang

More information

SP-CNN: A Scalable and Programmable CNN-based Accelerator. Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay

SP-CNN: A Scalable and Programmable CNN-based Accelerator. Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay SP-CNN: A Scalable and Programmable CNN-based Accelerator Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay Motivation Power is a first-order design constraint, especially for embedded devices. Certain

More information

Developing Monovalent Ion Parameters for the Optimal Point Charge (OPC) Water Model. John Dood Hope College

Developing Monovalent Ion Parameters for the Optimal Point Charge (OPC) Water Model. John Dood Hope College Developing Monovalent Ion Parameters for the Optimal Point Charge (OPC) Water Model John Dood Hope College What are MD simulations? Model and predict the structure and dynamics of large macromolecules.

More information

Molecular Dynamics Simulations

Molecular Dynamics Simulations Molecular Dynamics Simulations Dr. Kasra Momeni www.knanosys.com Outline Long-range Interactions Ewald Sum Fast Multipole Method Spherically Truncated Coulombic Potential Speeding up Calculations SPaSM

More information

Molecular Dynamics Simulation of a Biomolecule with High Speed, Low Power and Accuracy Using GPU-Accelerated TSUBAME2.

Molecular Dynamics Simulation of a Biomolecule with High Speed, Low Power and Accuracy Using GPU-Accelerated TSUBAME2. APSIPA ASC 2011 Xi an Molecular Dynamics Simulation of a Biomolecule with High Speed, Low Power and Accuracy Using GPU-Accelerated TSUBAME2.0 Supercomputer Shiqiao Du, Takuro Udagawa, Toshio Endo and Masakazu

More information

Hyeyoung Shin a, Tod A. Pascal ab, William A. Goddard III abc*, and Hyungjun Kim a* Korea

Hyeyoung Shin a, Tod A. Pascal ab, William A. Goddard III abc*, and Hyungjun Kim a* Korea The Scaled Effective Solvent Method for Predicting the Equilibrium Ensemble of Structures with Analysis of Thermodynamic Properties of Amorphous Polyethylene Glycol-Water Mixtures Hyeyoung Shin a, Tod

More information

Introduction to Benchmark Test for Multi-scale Computational Materials Software

Introduction to Benchmark Test for Multi-scale Computational Materials Software Introduction to Benchmark Test for Multi-scale Computational Materials Software Shun Xu*, Jian Zhang, Zhong Jin xushun@sccas.cn Computer Network Information Center Chinese Academy of Sciences (IPCC member)

More information

Research on GPU-accelerated algorithm in 3D finite difference neutron diffusion calculation method

Research on GPU-accelerated algorithm in 3D finite difference neutron diffusion calculation method NUCLEAR SCIENCE AND TECHNIQUES 25, 0501 (14) Research on GPU-accelerated algorithm in 3D finite difference neutron diffusion calculation method XU Qi ( 徐琪 ), 1, YU Gang-Lin ( 余纲林 ), 1 WANG Kan ( 王侃 ),

More information

Computation of non-bonded interactions: Part 1

Computation of non-bonded interactions: Part 1 Computation of non-bonded interactions: Part 1 Computation of non-local (non-bonded) interactions in the potential function E p non-local scales with the number of atoms as N. For the large molecular systems

More information

Klaus Schulten Department of Physics and Theoretical and Computational Biophysics Group University of Illinois at Urbana-Champaign

Klaus Schulten Department of Physics and Theoretical and Computational Biophysics Group University of Illinois at Urbana-Champaign Klaus Schulten Department of Physics and Theoretical and Computational Biophysics Group University of Illinois at Urbana-Champaign GTC, San Jose Convention Center, CA Sept. 20 23, 2010 GPU and the Computational

More information

Practical Combustion Kinetics with CUDA

Practical Combustion Kinetics with CUDA Funded by: U.S. Department of Energy Vehicle Technologies Program Program Manager: Gurpreet Singh & Leo Breton Practical Combustion Kinetics with CUDA GPU Technology Conference March 20, 2015 Russell Whitesides

More information

k θ (θ θ 0 ) 2 angles r i j r i j

k θ (θ θ 0 ) 2 angles r i j r i j 1 Force fields 1.1 Introduction The term force field is slightly misleading, since it refers to the parameters of the potential used to calculate the forces (via gradient) in molecular dynamics simulations.

More information

Direct Self-Consistent Field Computations on GPU Clusters

Direct Self-Consistent Field Computations on GPU Clusters Direct Self-Consistent Field Computations on GPU Clusters Guochun Shi, Volodymyr Kindratenko National Center for Supercomputing Applications University of Illinois at UrbanaChampaign Ivan Ufimtsev, Todd

More information

A CUDA Solver for Helmholtz Equation

A CUDA Solver for Helmholtz Equation Journal of Computational Information Systems 11: 24 (2015) 7805 7812 Available at http://www.jofcis.com A CUDA Solver for Helmholtz Equation Mingming REN 1,2,, Xiaoguang LIU 1,2, Gang WANG 1,2 1 College

More information

The Molecular Dynamics Method

The Molecular Dynamics Method H-bond energy (kcal/mol) - 4.0 The Molecular Dynamics Method Fibronectin III_1, a mechanical protein that glues cells together in wound healing and in preventing tumor metastasis 0 ATPase, a molecular

More information

Extending Parallel Scalability of LAMMPS and Multiscale Reactive Molecular Simulations

Extending Parallel Scalability of LAMMPS and Multiscale Reactive Molecular Simulations September 18, 2012 Extending Parallel Scalability of LAMMPS and Multiscale Reactive Molecular Simulations Yuxing Peng, Chris Knight, Philip Blood, Lonnie Crosby, and Gregory A. Voth Outline Proton Solvation

More information

DISCRETE TUTORIAL. Agustí Emperador. Institute for Research in Biomedicine, Barcelona APPLICATION OF DISCRETE TO FLEXIBLE PROTEIN-PROTEIN DOCKING:

DISCRETE TUTORIAL. Agustí Emperador. Institute for Research in Biomedicine, Barcelona APPLICATION OF DISCRETE TO FLEXIBLE PROTEIN-PROTEIN DOCKING: DISCRETE TUTORIAL Agustí Emperador Institute for Research in Biomedicine, Barcelona APPLICATION OF DISCRETE TO FLEXIBLE PROTEIN-PROTEIN DOCKING: STRUCTURAL REFINEMENT OF DOCKING CONFORMATIONS Emperador

More information

Molecular Dynamics. Molecules in motion

Molecular Dynamics. Molecules in motion Molecular Dynamics Molecules in motion 1 Molecules in mo1on Molecules are not sta1c, but move all the 1me Source: h9p://en.wikipedia.org/wiki/kine1c_theory 2 Gasses, liquids and solids Gasses, liquids

More information

Anton, a Special-Purpose Machine for Molecular Dynamics Simulation

Anton, a Special-Purpose Machine for Molecular Dynamics Simulation Anton, a Special-Purpose Machine for Molecular Dynamics Simulation David E. Shaw, Martin M. Deneroff, Ron O. Dror, Jeffrey S. Kuskin, Richard H. Larson, John K. Salmon, Cliff Young, Brannon Batson, Kevin

More information

Why study protein dynamics?

Why study protein dynamics? Why study protein dynamics? Protein flexibility is crucial for function. One average structure is not enough. Proteins constantly sample configurational space. Transport - binding and moving molecules

More information

Petascale Quantum Simulations of Nano Systems and Biomolecules

Petascale Quantum Simulations of Nano Systems and Biomolecules Petascale Quantum Simulations of Nano Systems and Biomolecules Emil Briggs North Carolina State University 1. Outline of real-space Multigrid (RMG) 2. Scalability and hybrid/threaded models 3. GPU acceleration

More information

Alchemical free energy calculations in OpenMM

Alchemical free energy calculations in OpenMM Alchemical free energy calculations in OpenMM Lee-Ping Wang Stanford Department of Chemistry OpenMM Workshop, Stanford University September 7, 2012 Special thanks to: John Chodera, Morgan Lawrenz Outline

More information

SPECIAL PROJECT PROGRESS REPORT

SPECIAL PROJECT PROGRESS REPORT SPECIAL PROJECT PROGRESS REPORT Progress Reports should be 2 to 10 pages in length, depending on importance of the project. All the following mandatory information needs to be provided. Reporting year

More information

Anton: A Specialized ASIC for Molecular Dynamics

Anton: A Specialized ASIC for Molecular Dynamics Anton: A Specialized ASIC for Molecular Dynamics Martin M. Deneroff, David E. Shaw, Ron O. Dror, Jeffrey S. Kuskin, Richard H. Larson, John K. Salmon, and Cliff Young D. E. Shaw Research Marty.Deneroff@DEShawResearch.com

More information

accelerated Molecular Dynamics (amd) Tutorial Levi Pierce 2012 NBCR Summer InsAtute

accelerated Molecular Dynamics (amd) Tutorial Levi Pierce 2012 NBCR Summer InsAtute accelerated Molecular Dynamics (amd) Tutorial Levi Pierce 2012 NBCR Summer InsAtute Time Scales Accessible with Molecular Dynamics Pierce, L.C.T.; Salomon- Ferrer, R.; de Oliveira C.A.; McCammon, J.A.;

More information

Computational Molecular Biophysics. Computational Biophysics, GRS Jülich SS 2013

Computational Molecular Biophysics. Computational Biophysics, GRS Jülich SS 2013 Computational Molecular Biophysics Computational Biophysics, GRS Jülich SS 2013 Computational Considerations Number of terms for different energy contributions / E el ~N ~N ~N ~N 2 Computational cost for

More information

MOLECULAR DYNAMICS SIMULATIONS OF BIOMOLECULES: Long-Range Electrostatic Effects

MOLECULAR DYNAMICS SIMULATIONS OF BIOMOLECULES: Long-Range Electrostatic Effects Annu. Rev. Biophys. Biomol. Struct. 1999. 28:155 79 Copyright c 1999 by Annual Reviews. All rights reserved MOLECULAR DYNAMICS SIMULATIONS OF BIOMOLECULES: Long-Range Electrostatic Effects Celeste Sagui

More information

Advanced sampling. fluids of strongly orientation-dependent interactions (e.g., dipoles, hydrogen bonds)

Advanced sampling. fluids of strongly orientation-dependent interactions (e.g., dipoles, hydrogen bonds) Advanced sampling ChE210D Today's lecture: methods for facilitating equilibration and sampling in complex, frustrated, or slow-evolving systems Difficult-to-simulate systems Practically speaking, one is

More information

Advanced Molecular Dynamics

Advanced Molecular Dynamics Advanced Molecular Dynamics Introduction May 2, 2017 Who am I? I am an associate professor at Theoretical Physics Topics I work on: Algorithms for (parallel) molecular simulations including GPU acceleration

More information

Molecular Dynamics Investigation of the ω-current in the Kv1.2 Voltage Sensor Domains

Molecular Dynamics Investigation of the ω-current in the Kv1.2 Voltage Sensor Domains Molecular Dynamics Investigation of the ω-current in the Kv1.2 Voltage Sensor Domains Fatemeh Khalili-Araghi, Emad Tajkhorshid, Benoît Roux, and Klaus Schulten Department of Physics, Department of Biochemistry,

More information

Performance Analysis of Lattice QCD Application with APGAS Programming Model

Performance Analysis of Lattice QCD Application with APGAS Programming Model Performance Analysis of Lattice QCD Application with APGAS Programming Model Koichi Shirahata 1, Jun Doi 2, Mikio Takeuchi 2 1: Tokyo Institute of Technology 2: IBM Research - Tokyo Programming Models

More information

Investigation of an Unusual Phase Transition Freezing on heating of liquid solution

Investigation of an Unusual Phase Transition Freezing on heating of liquid solution Investigation of an Unusual Phase Transition Freezing on heating of liquid solution Calin Gabriel Floare National Institute for R&D of Isotopic and Molecular Technologies, Cluj-Napoca, Romania Max von

More information

Introduction to Computer Simulations of Soft Matter Methodologies and Applications Boulder July, 19-20, 2012

Introduction to Computer Simulations of Soft Matter Methodologies and Applications Boulder July, 19-20, 2012 Introduction to Computer Simulations of Soft Matter Methodologies and Applications Boulder July, 19-20, 2012 K. Kremer Max Planck Institute for Polymer Research, Mainz Overview Simulations, general considerations

More information

Lattice Boltzmann simulations on heterogeneous CPU-GPU clusters

Lattice Boltzmann simulations on heterogeneous CPU-GPU clusters Lattice Boltzmann simulations on heterogeneous CPU-GPU clusters H. Köstler 2nd International Symposium Computer Simulations on GPU Freudenstadt, 29.05.2013 1 Contents Motivation walberla software concepts

More information

Claude Tadonki. MINES ParisTech PSL Research University Centre de Recherche Informatique

Claude Tadonki. MINES ParisTech PSL Research University Centre de Recherche Informatique Claude Tadonki MINES ParisTech PSL Research University Centre de Recherche Informatique claude.tadonki@mines-paristech.fr Monthly CRI Seminar MINES ParisTech - CRI June 06, 2016, Fontainebleau (France)

More information

Massive Parallelization of First Principles Molecular Dynamics Code

Massive Parallelization of First Principles Molecular Dynamics Code Massive Parallelization of First Principles Molecular Dynamics Code V Hidemi Komatsu V Takahiro Yamasaki V Shin-ichi Ichikawa (Manuscript received April 16, 2008) PHASE is a first principles molecular

More information

MD Thermodynamics. Lecture 12 3/26/18. Harvard SEAS AP 275 Atomistic Modeling of Materials Boris Kozinsky

MD Thermodynamics. Lecture 12 3/26/18. Harvard SEAS AP 275 Atomistic Modeling of Materials Boris Kozinsky MD Thermodynamics Lecture 1 3/6/18 1 Molecular dynamics The force depends on positions only (not velocities) Total energy is conserved (micro canonical evolution) Newton s equations of motion (second order

More information

Extending the generality of molecular dynamics simulations on a special-purpose machine

Extending the generality of molecular dynamics simulations on a special-purpose machine Extending the generality of molecular dynamics simulations on a special-purpose machine Daniele P. Scarpazza, Douglas J. Ierardi, Adam K. Lerer, Kenneth M. Mackenzie, Albert C. Pan, Joseph A. Bank, Edmond

More information

Towards fast and accurate binding affinity. prediction with pmemdgti: an efficient. implementation of GPU-accelerated. Thermodynamic Integration

Towards fast and accurate binding affinity. prediction with pmemdgti: an efficient. implementation of GPU-accelerated. Thermodynamic Integration Towards fast and accurate binding affinity prediction with pmemdgti: an efficient implementation of GPU-accelerated Thermodynamic Integration Tai-Sung Lee,, Yuan Hu, Brad Sherborne, Zhuyan Guo, and Darrin

More information

Accelerated Quantum Molecular Dynamics

Accelerated Quantum Molecular Dynamics Accelerated Quantum Molecular Dynamics Enrique Martinez, Christian Negre, Marc J. Cawkwell, Danny Perez, Arthur F. Voter and Anders M. N. Niklasson Outline Quantum MD Current approaches Challenges Extended

More information

MARTINI simulation details

MARTINI simulation details S1 Appendix MARTINI simulation details MARTINI simulation initialization and equilibration In this section, we describe the initialization of simulations from Main Text section Residue-based coarsegrained

More information

Methods of Computer Simulation. Molecular Dynamics and Monte Carlo

Methods of Computer Simulation. Molecular Dynamics and Monte Carlo Molecular Dynamics Time is of the essence in biological processes therefore how do we understand time-dependent processes at the molecular level? How do we do this experimentally? How do we do this computationally?

More information

A Brief Guide to All Atom/Coarse Grain Simulations 1.1

A Brief Guide to All Atom/Coarse Grain Simulations 1.1 A Brief Guide to All Atom/Coarse Grain Simulations 1.1 Contents 1. Introduction 2. AACG simulations 2.1. Capabilities and limitations 2.2. AACG commands 2.3. Visualization and analysis 2.4. AACG simulation

More information

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics SPARSE SOLVERS FOR THE POISSON EQUATION Margreet Nool CWI, Multiscale Dynamics November 9, 2015 OUTLINE OF THIS TALK 1 FISHPACK, LAPACK, PARDISO 2 SYSTEM OVERVIEW OF CARTESIUS 3 POISSON EQUATION 4 SOLVERS

More information

Computational Chemistry - MD Simulations

Computational Chemistry - MD Simulations Computational Chemistry - MD Simulations P. Ojeda-May pedro.ojeda-may@umu.se Department of Chemistry/HPC2N, Umeå University, 901 87, Sweden. May 2, 2017 Table of contents 1 Basics on MD simulations Accelerated

More information