Multiscale simulations of complex fluid rheology

Similar documents
Randomly Triangulated Surfaces as Models for Fluid and Crystalline Membranes. G. Gompper Institut für Festkörperforschung, Forschungszentrum Jülich

Lattice Boltzmann simulations on heterogeneous CPU-GPU clusters

GPU Acceleration of Cutoff Pair Potentials for Molecular Modeling Applications

Mesoscale fluid simulation of colloidal systems

A microsecond a day keeps the doctor away: Efficient GPU Molecular Dynamics with GROMACS

Parallelization of Molecular Dynamics (with focus on Gromacs) SeSE 2014 p.1/29

DISCRETE TUTORIAL. Agustí Emperador. Institute for Research in Biomedicine, Barcelona APPLICATION OF DISCRETE TO FLEXIBLE PROTEIN-PROTEIN DOCKING:

COMPARISON OF CPU AND GPU IMPLEMENTATIONS OF THE LATTICE BOLTZMANN METHOD

Mesoscale Modeling of Blood Flow: From Single Cells to Blood Rheology

Multiphase Flow Simulations in Inclined Tubes with Lattice Boltzmann Method on GPU

Dissipative Particle Dynamics: Foundation, Evolution and Applications

Performance Evaluation of Scientific Applications on POWER8

LAMMPS Performance Benchmark on VSC-1 and VSC-2

Smoothed Dissipative Particle Dynamics: theory and applications to complex fluids

File ISM02. Dynamics of Soft Matter

A Framework for Hybrid Parallel Flow Simulations with a Trillion Cells in Complex Geometries

Dresden, September 20-24, 2010 by Hartmut Löwen

Directed Assembly of Functionalized Nanoparticles with Amphiphilic Diblock Copolymers. Contents

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

On Portability, Performance and Scalability of a MPI OpenCL Lattice Boltzmann Code

The Molecular Dynamics Method

Hybrid Atomistic-Continuum Methods for Dense Liquids

Parallel Simulations of Self-propelled Microorganisms

Direct Self-Consistent Field Computations on GPU Clusters

Welcome to MCS 572. content and organization expectations of the course. definition and classification

Efficient Parallelization of Molecular Dynamics Simulations on Hybrid CPU/GPU Supercoputers

Direct Modeling for Computational Fluid Dynamics

ATOMISTIC/CONTINUUM MULTISCALE COUPLING

arxiv:cond-mat/ v1 [cond-mat.stat-mech] 11 Dec 2002

The Fast Multipole Method in molecular dynamics

MD simulation of methane in nanochannels

Coupling Atomistic and Continuum Hydrodynamics

RWTH Aachen University

MOLECULAR DYNAMIC SIMULATION OF WATER VAPOR INTERACTION WITH VARIOUS TYPES OF PORES USING HYBRID COMPUTING STRUCTURES

CE 530 Molecular Simulation

Multiscale modeling of active fluids: selfpropellers and molecular motors. I. Pagonabarraga University of Barcelona

Supporting Information Soft Nanoparticles: Nano Ionic Networks of Associated Ionic Polymers

Introduction to Benchmark Test for Multi-scale Computational Materials Software

Q1. (a) Define the term activation energy for a chemical reaction. (2)

An electrokinetic LB based model for ion transport and macromolecular electrophoresis

Scalable Hybrid Programming and Performance for SuperLU Sparse Direct Solver

Theory, Algorithms and Applications of Dissipative Particle Dynamics. Sept , Shanghai

An Event-Driven Hybrid Molecular Dynamics and Direct Simulation Monte Carlo Algorithm. Abstract

Performance of WRF using UPC

Extending Parallel Scalability of LAMMPS and Multiscale Reactive Molecular Simulations

Nucleation of nanoparticles in a coarse grained fluid using OpenCL

N-body Simulations. On GPU Clusters

Radiation Effect on MHD Casson Fluid Flow over a Power-Law Stretching Sheet with Chemical Reaction

Petascale Quantum Simulations of Nano Systems and Biomolecules

Coarse-grained Models for Oligomer-grafted Silica Nanoparticles

Performance Evaluation of MPI on Weather and Hydrological Models

Pore Scale Analysis of Oil Shale/Sands Pyrolysis

Coupling of nanoparticle dynamics to polymer. center-of-mass motion in semidilute polymer. solutions

The Lattice Boltzmann Method for Laminar and Turbulent Channel Flows

An introduction to the lattice Boltzmann method for multiphase flows with interfacially active components. Roar Skartlien 2012

arxiv: v1 [cond-mat.soft] 8 Mar 2018

What is the role of simulation in nanoscience research?

MD Thermodynamics. Lecture 12 3/26/18. Harvard SEAS AP 275 Atomistic Modeling of Materials Boris Kozinsky

Potential Energy (hyper)surface

Particle-Simulation Methods for Fluid Dynamics

BUDE. A General Purpose Molecular Docking Program Using OpenCL. Richard B Sessions

Faster Kinetics: Accelerate Your Finite-Rate Combustion Simulation with GPUs

Coarse-graining, applications and mesoscopic molecular dynamics

Heterogeneous programming for hybrid CPU-GPU systems: Lessons learned from computational chemistry

Model Studies on Slag-Metal Entrainment in Gas Stirred Ladles

Introduction to numerical computations on the GPU

Outline. Advances in STAR-CCM+ DEM models for simulating deformation, breakage, and flow of solids

Hybrid CPU/GPU Acceleration of Detection of 2-SNP Epistatic Interactions in GWAS

Kinetic/Fluid micro-macro numerical scheme for Vlasov-Poisson-BGK equation using particles

Tuning And Understanding MILC Performance In Cray XK6 GPU Clusters. Mike Showerman, Guochun Shi Steven Gottlieb

A Coupling Tool for Parallel Molecular Dynamics Continuum Simulations

Optics and Spectroscopy

Performance of the fusion code GYRO on three four generations of Crays. Mark Fahey University of Tennessee, Knoxville

Using AmgX to accelerate a PETSc-based immersed-boundary method code

Modeling of colloidal gels

MONTE CARLO NEUTRON TRANSPORT SIMULATING NUCLEAR REACTIONS ONE NEUTRON AT A TIME Tony Scudiero NVIDIA

Coupling of Multi fidelity Models Applications to PNP cdft and local nonlocal Poisson equations

FLOW ASSURANCE: DROP COALESCENCE IN THE PRESENCE OF SURFACTANTS

Interface Roughening in a Hydrodynamic Lattice- Gas Model with Surfactant

Parallel Performance Studies for a Numerical Simulator of Atomic Layer Deposition Michael J. Reid

Non equilibrium thermodynamics: foundations, scope, and extension to the meso scale. Miguel Rubi

Smoothed Dissipative Particle Dynamics Model for Predicting Self-Assembled Nano-Cellulose Fibre Structures

The Effect of Model Internal Flexibility Upon NEMD Simulations of Viscosity

Simulation of Lid-driven Cavity Flow by Parallel Implementation of Lattice Boltzmann Method on GPUs

Coupling of nanoparticle dynamics to polymer. center-of-mass motion in semidilute polymer. solutions

Linking the Continuous and the Discrete

Scalable and Power-Efficient Data Mining Kernels

Hydrodynamics in Espresso. Kai Grass 1

Lecture 5: Macromolecules, polymers and DNA

Technologies VII. Alternative Lithographic PROCEEDINGS OF SPIE. Douglas J. Resnick Christopher Bencher. Sponsored by. Cosponsored by.

Massively parallel molecular-continuum simulations with the macro-micro-coupling tool Neumann, P.; Harting, J.D.R.

Multiscale Methods for Hydrodynamics of Complex Fluids

arxiv:comp-gas/ v1 28 Apr 1993

Coarse-graining limits in open and wall-bounded dissipative particle dynamics systems

Simulation of a 3D Flow-Focusing Capillary-Based Droplet Generator

Accelerating Three-Body Molecular Dynamics Potentials Using NVIDIA Tesla K20X GPUs. GE Global Research Masako Yamada

Using a CUDA-Accelerated PGAS Model on a GPU Cluster for Bioinformatics

An Efficient FETI Implementation on Distributed Shared Memory Machines with Independent Numbers of Subdomains and Processors

Supplementary Information for: Controlling Cellular Uptake of Nanoparticles with ph-sensitive Polymers

Mesoscopic Simulation of Aggregate Structure and Stability of Heavy Crude Oil by GPU Accelerated DPD

Transcription:

Multiscale simulations of complex fluid rheology Michael P. Howard, Athanassios Z. Panagiotopoulos Department of Chemical and Biological Engineering, Princeton University Arash Nikoubashman Institute of Physics, Johannes Gutenberg University Mainz 18 May 2017 mphoward@princeton.edu

Complex fluids are everywhere 2

An industrial example: enhanced oil recovery Primary recovery: pressure drives the oil from the well (10%) Secondary recovery: water or gas displaces oil (~30%) Tertiary (enhanced) oil recovery by gas, water, chemical injection Surfactants, polymers, nanoparticles, etc. What is the best formulation? What fundamental physics control transport of these fluids? Computer simulations can allow for rapid exploration of the design space for complex fluids 3

Multiscale simulation model s ms resolution Constitutive models time scale µs ns ps Atomistic models Coarse-grained & mesoscale models speed fs nm µm mm length scale 4

Multiscale simulation model bond stretch angle bend dihedral twist dispersion forces, excluded volume, electrostatics solvent effects 5

Multiparticle collision dynamics Mesoscale simulation method that coarse-grains the solvent while preserving hydrodynamics. 1 Alternate solvent particle streaming with collisions. Collisions are momentum- and energy-conserving. An H-theorem exists, and the correct Maxwell-Boltzmann distribution of velocities is obtained. The Navier- Stokes momentum balance is satisfied. 1 A. Malevanets and R. Kapral. J. Chem. Phys. 110, 8605-8613 (1999). 6

Multiparticle collision dynamics Polymers are coupled to the solvent during the collision step. 2 MPCD MD Δt MD Δt stream collide Δt stream collide Rigid objects are coupled during the streaming step. 3 T F 2 A. Malevanets and J.M. Yeomans. Europhys. Lett. 52, 231-237 (2000). 3 G. Gompper, T. Ihle, D.M. Kroll, and R.G. Winkler. Advanced Computer Simulation Approaches for Soft Matter Sciences III, Advances in Polymer Science Vol. 221 (Springer, 2009), p. 1-87. 7

Multiparticle collision dynamics 4 A. Nikoubashman, N.A. Mahynski, A.H. Pirayandeh, and A.Z. Panagiotopoulos. J. Chem. Phys. 140, 094903 (2014). 5 M.P. Howard, A.Z. Panagiotopoulos, and A. Nikoubashman, J. Chem. Phys. 142, 224908 (2015). 8

Why Blue Waters? Typical MPCD simulations require many particles with simple interactions. ( 10 particles per cell ) x ( 100 3-400 3 cells ) = 10 640 million particles MPCD readily lends itself to parallelization because of its particle- and cellbased nature: 1. Accelerators (e.g., GPU) within a node. 2. Domain decomposition to many nodes. For molecular dynamics, GPU acceleration can significantly increase the scientific throughput by as much as an order of magnitude. Blue Waters is the only NSF-funded system currently giving access to GPU resources at massive scale! 9

Implementation HOOMD-blue simulation package 6,7 for design philosophy and flexible user interface. import hoomd hoomd.context.initialize() from hoomd import mpcd # initialize an empty hoomd system first box = hoomd.data.boxdim(l=200.) hoomd.init.read_snapshot(hoomd.data.make_snapshot(n=0, box=box)) # then initialize the MPCD system randomly s = mpcd.init.make_random(n=int(10*box.get_volume()), kt=1.0, seed=7) s.sorter.set_period(period=100) # use the SRD collision rule ig = mpcd.integrator(dt=0.1, period=1) mpcd.collide.srd(seed=1, period=1, angle=130.) # run for 5000 steps hoomd.run(5000) HOOMD-blue is available open-source for use on Blue Waters: http://glotzerlab.engin.umich.edu/hoomd-blue 6 J.A. Anderson, C.D. Lorenz, and A. Travesset. J. Comput. Phys. 227, 5342-5359 (2008). 7 J. Glaser et al. Comput. Phys. Commun. 192, 97-107 (2015). 10

Implementation: cell list particles N c 1 0 4 8 5 2 7 6 3 9 atomic write cell index 1 4 2 7 3 9 6 0 5 8 2 5 2 0 1 Improve performance by periodically sorting particles. compact 1 4 2 7 3 9 6 0 5 8 renumber 0 1 2 3 4 5 6 7 8 9 11

Implementation: domain decomposition HOOMD-blue supports non-uniform domain boundaries. To simplify integration with molecular dynamics code, we adopt the same decomposition. domain boundary simulation box cells covered domain Any cells overlapping between domains require communication for computing cell-level averages, like the velocity or kinetic energy. All MPI buffers are packed and unpacked on the GPU. Some performance can be gained from using CUDA-aware MPI when GPUDirect RDMA is available. 12

Implementation: CUDA optimizations Wide parallelism w threads per cell 7 3 2 9 6 4 7 3 2 9 6 4 strided evaluation Runtime autotuning Significant performance variation from size of the thread-block that is architecture and system-size dependent. block size w K20x 448 4 P100 224 16 2 2 1 1 3 3 6 warp reduction Solution: use periodic runtime tuning with CUDA event timings to scan through and select optimal parameters. m_tuner >begin(); gpu::cell_thermo(..., m_tuner >getparam()); m_tuner >end(); Only a small fraction of time is spent running at non-optimal parameters for most kernels. 13

Implementation: random number generation Need efficient random number generation for the collision step. One PRNG state is created per-thread and per-kernel using a cryptographic hash. 8 Micro-streams of random numbers are then drawn. global cell index timestep seed Saru hash (45 instr.) draw (13 instr.) draw The additional instructions needed to prepare the PRNG state can be more efficient than loading a state from memory on the GPU. Additional benefit: significantly reduced cell-level MPI communication since no synchronization of random numbers is required. 8 C.L. Phillips, J.A. Anderson, and S.C. Glotzer. J. Comput. Phys. 230, 7191-7201 (2011). 14

Performance on Blue Waters Strong scaling tested for L = 200 and L = 400 cubic boxes with 10 particles / cell (80 million and 640 million particles). time steps per second 100 10 1 GPU L/a = 200 L/a = 400 CPU L/a = 200 L/a = 400 time [ms / step] 100 10 L/a = 400 total core particle comm. cell comm. 1 2 4 8 16 32 64 128 256 512 1024 number of nodes 32 64 128 256 512 1024 number of nodes CPU benchmarks (16 processes per XE node) show excellent scaling. GPU benchmarks (1 process per XK node) show good scaling, with some loss of performance at high node counts. GPU profile shows that performance is bottlenecked by cell communication. 15

Future plans Performance enhancements 1. Overlap outer cell communication with inner cell computations. 2. Test a mixed-precision model for floating-point particle and cell data. Feature enhancements 1. Implement additional MPCD collision rules for a multiphase fluid. 2. Couple particles to solid bounding walls. Scientific application: Compare MPCD multiphase fluid to MD simulations, and study droplet physics in confined complex fluid. All code is available online! https://bitbucket.org/glotzer/hoomd-blue 16