Preskoči na sadržaj

Molecular dynamics simulation for exascale supercomputing era: scientific research and software engineering challenges

Vedran Miletić 😎, Matea Turalija 😎

😎 Group for Applications and Services on Exascale Research Infrastructure (GASERI), Faculty of Informatics and Digital Technologies (FIDIT), University of Rijeka

Invited lecture at the Computational Chemistry Day in Zagreb, Croatia, 16th of September, 2023.

First off, thanks to the organizers for the invitation

Computational Chemistry Day 2023 Logo

Image source: Computational Chemistry Day Logo

Faculty of Informatics and Digital Technologies (FIDIT)

  • founded in 2008. as Department of Informatics at University of Rijeka
  • became Faculty of Informatics and Digital Technologies in 2022.
  • several laboratories and research groups; scientific interests:
    • computer vision, pattern recognition
    • natural language processing, machine translation
    • e-learning, digital transformation
    • scientific software parallelization, high performance computing

FIDIT logo bg right:35% 95%

Group for Applications and Services on Exascale Research Infrastructure (GASERI)

  • focus: optimization of computational biochemistry applications for running on modern exascale supercomputers
  • vision: high performance algorithms readily available
    • for academic and industrial use
    • in state-of-the-art open-source software
  • members: Dr. Vedran Miletić (PI), Matea Turalija (PhD student), Milan Petrović (teaching assistant)


How did we end up here, in computational chemistry?

  • Prof. Dr. Branko Mikac's optical network simulation group at FER, Department of Telecommunications
  • What to do after finishing the Ph.D. thesis? 🤔
  • postdoc in Prof. Frauke Gräter's Molecular Biomechanics (MBM) group at Heidelberg Institute for Theoretical Studies (HITS)
    • collaboration with GROMACS molecular dynamics simulation software developers from KTH, MPI-NAT, UMich, CU, and others

High performance computing

  • a supercomputer is a computer with a high level of performance as compared to a general-purpose computer
    • also called high performance computer (HPC)
  • measure: floating-point operations per second (FLOPS)
    • PC -> teraFLOPS; Bura -> 100 teraFLOPS, Supek -> 1,2 petaFLOPS
    • modern HPC -> 1 do 10 petaFLOPS, top 1194 petaFLOPS (~1,2 exaFLOPS)
    • future HPC -> 2+ exaFLOPS
  • nearly exponential growth of FLOPS over time (source: Wikimedia Commons File:Supercomputers-history.svg)

bg 80% Computing power of the top 1 supercomputer each year, measured in FLOPS

Languages used in scientific software development are also evolving to utilize the new hardware

  • C++: C++11 (C++0x), C++14, C++17, C++20, C++23
    • Parallel standard library
    • Task blocks
  • Fortran: Fortran 2008, Fortran 2018 (2015), Fortran 2023
    • Coarray Fortran
    • DO CONCURRENT for loops
    • CONTIGUOUS storage layout
  • New languages: Python, Julia, Rust, ...

More heterogeneous architectures require complex programming models


Parallelization for a large HPC requires domain decomposition

DD cells

Image source for both figures: Parallelization (GROMACS Manual)

bg right 70% DD triclinical

Domain decomposition and particle mesh Ewald

MPMD PME width:950px

Image source for both figures: Parallelization (GROMACS Manual)

This approachs still does not scale indefinitely. Is there an alternative?


The particle mesh Ewald method (PME, currently state of the art in molecular simulation) does not scale to large core counts as it suffers from a communication bottleneck, and does not treat titratable sites efficiently.

The fast multipole method (FMM) will enable an efficient calculation of long-range interactions on massively parallel exascale computers, including alternative charge distributions representing various forms of titratable sites.

SPPEXA Projects - Phase 2 (2016 - 2018)


FlexFMM Logo height:180px

Image source: Max Planck Institute for Multidisciplinary Sciences FlexFMM

Our GROMACS developments: generalized FMM

  • molecular dynamics simulations are periodic with various simulation box types: cubic, rhombic dodecahedron; present design and implementation of the fast multipole method supports only cubic boxes
    • many useful applications (materials, interfaces) fit well into rectangular cuboid boxes, not cubic -> Matea's PhD thesis research

Cuboid height:200px

Image source: Wikimedia Commons File:Cuboid no label.svg

  • it is possible to also support rhombic dodecahedron: ~30% less volume => ~30% less computation time per step required

Our GROMACS developments: NVIDIA BlueField

NVIDIA Logo height:100px

Image source: Wikimedia Commons File:NVIDIA logo.svg

Our GROMACS developments: weight factor expressions and generalized flow simulation (1/4)

100 ps molecular dynamics simulation of water. width:700px

Image source: Wikimedia Commons File:MD water.gif

Our GROMACS developments: weight factor expressions and generalized flow simulation (2/4)

  • Flow is a movement of solvent atoms
  • Pulling all solvent atoms works when no other molecules except water are present in the simulation
  • Just a slice of solvent molecules should be pulled to allow the solvent atoms to interact with the biomolecule(s) without being "dragged away"

Our GROMACS developments: weight factor expressions and generalized flow simulation (3/4)

A rectangular water box with ~32,000 water molecules is shown. width:900px

Image source: Biophys. J. 116(6), 621–632 (2019). doi:10.1016/j.bpj.2018.12.025

Our GROMACS developments: weight factor expressions and generalized flow simulation (4/4)

  • Atom weight = dynamic weight factor computed from the expression x weight factor specified in the parameters file x atom mass-derived weight factor
  • Dynamic weight factor (and atom weight) recomputed in each simulation step
  • Weight factor expression variables:
    • Atom position in 3D (x, y, z)
    • Atom velocity in 3D (vx, vy, vz)
  • Examples:
    • Atom weight factor is sum of squares of positions: x^2 + y^2 + z^2
    • Atom weight factor is a linear combination of velocities: 1.75 * vx + 1.5 * vy + 1.25 * vz

Our potential GROMACS developments

  • Monte Carlo (Davide Mercadante, University of Auckland)
    • many efforts over the years, none with broad acceptance
    • should be rethought, and then designed and implemented from scratch with exascale in mind
  • polarizable simulations using the classical Drude oscillator model (Justin Lemkul, Virginia Tech)
    • should be parallelized for multi-node execution
  • other drug design tools such as Random Acceleration Molecular Dynamics (Rebecca Wade, Heidelberg Institute for Theoretical Studies and Daria Kokh, Cancer Registry of Baden-Württemberg)

Affiliation changes (1/2)

  • starting October, Matea will be joining Faculty of Medicine, University of Rijeka
  • PhD topic staying as planned and presented here

Image source: Wikimedia Commons File:Medicinski fakultet Rijeka 0710 1.jpg

Faculty of Medicine bg right:60%

Affiliation changes (2/2)

Image source: Wikimedia Commons File:110716031-TUM.JPG

TUM Campus bg right:40%

Thank you for your attention


GASERI website:

Vedran bg left

Matea bg right

Author: Vedran Miletić, Matea Turalija