Create large-scale ML-driven multiscale simulation ensembles to study the interactions

Overview

MuMMI RAS v0.1

Released: Nov 16, 2021

MuMMI RAS is the application component of the MuMMI framework developed to create large-scale ML-driven multiscale simulation ensembles to study the interactions of RAS proteins and RAS-RAF protein complexes with lipid plasma membranes.

MuMMI framework was developed as part of the Pilot2 project of the Joint Design of Advanced Computing Solutions for Cancer funded jointly by the Department of Energy (DOE) and the National Cancer Institute (NCI).

The Pilot 2 project focuses on developing multiscale simulation models for understanding the interactions of the lipid plasma membrane with the RAS and RAF proteins. The broad computational tool development aims of this pilot are:

  • Developing scalable multi-scale molecular dynamics code that will automatically switch between phase field, coarse-grained and all-atom simulations.
  • Developing scalable machine learning and predictive models of molecular simulations to:
    • identify and quantify states from simulations
    • identify events from simulations that can automatically signal change of resolution between phase field, coarse-grained and all-atom simulations
    • aggregate information from the multi-resolution simulations to efficiently feedback to/from machine learning tools
  • Integrate sparse information from experiments with simulation data

MuMMI RAS defines the specific functionalities needed for the various components and scales of a target multiscale simulation. The application components need to define the scales, how to read the corresponding data, how to perform ML-based selection, how to run the simulations, how to perform analysis, and how to perform feedback. This code uses several utilities made available through "MuMMI Core".

Publications

MuMMI framework is described in the following publications.

  1. Bhatia et al. Generalizable Coordination of Large Multiscale Ensembles: Challenges and Learnings at Scale. In Proceedings of the ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, SC '21, Article No. 10, November 2021. doi:10.1145/3458817.3476210.

  2. Di Natale et al. A Massively Parallel Infrastructure for Adaptive Multiscale Simulations: Modeling RAS Initiation Pathway for Cancer. In Proceedings of the ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, SC '19, Article No. 57, November 2019. doi:10.1145/3295500.3356197.
    Best Paper at SC 2019.

  3. Ingólfsson et al. Machine Learning-driven Multiscale Modeling Reveals Lipid-Dependent Dynamics of RAS Signaling Protein. Proceedings of the National Academy of Sciences (PNAS), accepted, 2021. preprint.

  4. Reciprocal Coupling of Coarse-Grained and All-Atom scales. In preparation.

Installation

git clone https://github.com/mummi-framework/mummi-ras
cd mummi-ras
pip3 install .

export MUMMI_ROOT=/path/to/outputs
export MUMMI_CORE=/path/to/core/repo
export MUMMI_APP=/path/to/app/repo
export MUMMI_RESOURCES=/path/to/resources
The installaton process as described above installs the MuMMI framework. The simulation codes (gridsim2d, ddcMD, AMBER, GROMACS) are not included and are to be installed separately.
Spack installation. We are also working towards releasing the option of installing MuMMI and its dependencies through spack.

Authors and Acknowledgements

MuMMI was developed at Lawrence Livermore National Laboratory, in collaboration with Los Alamos National Laboratory, Oak Ridge National Laboratory, and International Business Machines. A list of main contributors is given below.

  • LLNL: Harsh Bhatia, Francesco Di Natale, Helgi I Ingólfsson, Joseph Y Moon, Xiaohua Zhang, Joseph R Chavez, Fikret Aydin, Tomas Oppelstrup, Timothy S Carpenter, Shiv Sundaram (previously LLNL), Gautham Dharuman (previously LLNL), Dong H Ahn, Stephen Herbein, Tom Scogland, Peer-Timo Bremer, and James N Glosli.

  • LANL: Chris Neale and Cesar Lopez

  • ORNL: Chris Stanley

  • IBM: Sara K Schumacher

MuMMI was funded by the Pilot2 project led by Dr. Fred Streitz (DOE) and Dr. Dwight Nissley (NIH). We acknowledge contributions from the entire Pilot 2 team.

This work was performed under the auspices of the U.S. Department of Energy (DOE) by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344, Los Alamos National Laboratory (LANL) under Contract DE-AC5206NA25396, and Oak Ridge National Laboratory under Contract DE-AC05-00OR22725.

Contact: Lawrence Livermore National Laboratory, 7000 East Avenue, Livermore, CA 94550.

Contributing

Contributions may be made through pull requests and/or issues on github.

License

MuMMI RAS is distributed under the terms of the MIT License.

Livermore Release Number: LLNL-CODE-827655

Comments
  • Are the trajectories in your publications publicly available?

    Are the trajectories in your publications publicly available?

    Hi, Congrats on the success, and huge thanks for making it open source. I wonder whether the trajectories in your publications are publicly available. Or are there any demo trajectories?

    I am a Ph.D. student at KAUST, using computer graphics to build and visualize mesoscale biology models, such as SARS-CoV-2 and bacteriophage T4. If possible, I (and my colleagues) would like to perform (multiscale, multi-representation, multi-granularity) visualization research on the trajectories you generated.

    Many thanks, Roden

    opened by RodenLuo 2
  • `flux` vs `slurm`

    `flux` vs `slurm`

    Hi,

    As flux is mentioned in the dependencies, is it possible to reproduce MuMMI RAS on a cluster that only has slurm?

    Workflow dependencies (e.g., python, flux, dynim, keras, etc.)

    Quoted from: https://github.com/mummi-framework/mummi-ras/blob/main/INSTALL.md

    Many thanks, Roden

    opened by RodenLuo 0
  • gridsim2d availability

    gridsim2d availability

    Hi, I wonder if the following code is available or not.

    gridsim2d: to be released shortly

    Quoted from: https://github.com/mummi-framework/mummi-ras/blob/main/INSTALL.md

    Thanks, Roden

    opened by RodenLuo 0
  • Patch for gromacs availability

    Patch for gromacs availability

    Hi, I wonder if the following patch is available or not.

    Note that we have a patch for gromacs installation for customization. To be open-sourced soon.

    Quoted from: https://github.com/mummi-framework/mummi-ras/blob/main/INSTALL.md

    Thanks, Roden

    opened by RodenLuo 0
  • Small scale test data for local deployment

    Small scale test data for local deployment

    Hi, I'm interested in deploying MuMMI on the KAUST IBEX cluster. It is mentioned in the installation doc that there is a small set of test data. Is it now publicly available? If not, is it possible for me to somehow access it so that I can perform a test run?

    Many thanks, Roden

    Again on lassen and on summit, we have created a small set of test data, which can be used to launch MuMMI at small scales. This (and the larger dataset) will be made public through NCI website. Until then, we can make this data available upon request.

    opened by RodenLuo 1
Releases(v1.0.0)
A simple machine learning package to cluster keywords in higher-level groups.

Simple Keyword Clusterer A simple machine learning package to cluster keywords in higher-level groups. Example: "Senior Frontend Engineer" -- "Fronte

Andrea D'Agostino 10 Dec 18, 2022
[DEPRECATED] Tensorflow wrapper for DataFrames on Apache Spark

TensorFrames (Deprecated) Note: TensorFrames is deprecated. You can use pandas UDF instead. Experimental TensorFlow binding for Scala and Apache Spark

Databricks 757 Dec 31, 2022
Class-imbalanced / Long-tailed ensemble learning in Python. Modular, flexible, and extensible

IMBENS: Class-imbalanced Ensemble Learning in Python Language: English | Chinese/中文 Links: Documentation | Gallery | PyPI | Changelog | Source | Downl

Zhining Liu 176 Jan 04, 2023
Fourier-Bayesian estimation of stochastic volatility models

fourier-bayesian-sv-estimation Fourier-Bayesian estimation of stochastic volatility models Code used to run the numerical examples of "Bayesian Approa

15 Jun 20, 2022
Book Item Based Collaborative Filtering

Book-Item-Based-Collaborative-Filtering Collaborative filtering methods are used

Şebnem 3 Jan 06, 2022
Hypernets: A General Automated Machine Learning framework to simplify the development of End-to-end AutoML toolkits in specific domains.

A General Automated Machine Learning framework to simplify the development of End-to-end AutoML toolkits in specific domains.

DataCanvas 216 Dec 23, 2022
Exemplary lightweight and ready-to-deploy machine learning project

Exemplary lightweight and ready-to-deploy machine learning project

snapADDY GmbH 6 Dec 20, 2022
CorrProxies - Optimizing Machine Learning Inference Queries with Correlative Proxy Models

CorrProxies - Optimizing Machine Learning Inference Queries with Correlative Proxy Models

ZhihuiYangCS 8 Jun 07, 2022
Implementation of linesearch Optimization Algorithms in Python

Nonlinear Optimization Algorithms During my time as Scientific Assistant at the Karlsruhe Institute of Technology (Germany) I implemented various Opti

Paul 3 Dec 06, 2022
Solve automatic numerical differentiation problems in one or more variables.

numdifftools The numdifftools library is a suite of tools written in _Python to solve automatic numerical differentiation problems in one or more vari

Per A. Brodtkorb 181 Dec 16, 2022
This project impelemented for midterm of the Machine Learning #Zoomcamp #Alexey Grigorev

MLProject_01 This project impelemented for midterm of the Machine Learning #Zoomcamp #Alexey Grigorev Context Dataset English question data set file F

Hadi Nakhi 1 Dec 18, 2021
Repositório para o #alurachallengedatascience1

1° Challenge de Dados - Alura A Alura Voz é uma empresa de telecomunicação que nos contratou para atuar como cientistas de dados na equipe de vendas.

Sthe Monica 16 Nov 10, 2022
Coursera Machine Learning - Python code

Coursera Machine Learning This repository contains python implementations of certain exercises from the course by Andrew Ng. For a number of assignmen

Jordi Warmenhoven 859 Dec 10, 2022
30 Days Of Machine Learning Using Pytorch

Objective of the repository is to learn and build machine learning models using Pytorch. 30DaysofML Using Pytorch

Mayur 119 Nov 24, 2022
This repo includes some graph-based CTR prediction models and other representative baselines.

Graph-based CTR prediction This is a repository designed for graph-based CTR prediction methods, it includes our graph-based CTR prediction methods: F

Big Data and Multi-modal Computing Group, CRIPAC 47 Dec 30, 2022
Forecast dynamically at scale with this unique package. pip install scalecast

🌄 Scalecast: Dynamic Forecasting at Scale About This package uses a scaleable forecasting approach in Python with common scikit-learn and statsmodels

Michael Keith 158 Jan 03, 2023
A handy tool for common machine learning models' hyper-parameter tuning.

Common machine learning models' hyperparameter tuning This repo is for a collection of hyper-parameter tuning for "common" machine learning models, in

Kevin Hu 2 Jan 27, 2022
Learn Machine Learning Algorithms by doing projects in Python and R Programming Language

Learn Machine Learning Algorithms by doing projects in Python and R Programming Language. This repo covers all aspect of Machine Learning Algorithms.

Ravi Chaubey 6 Oct 20, 2022
🎛 Distributed machine learning made simple.

🎛 lazycluster Distributed machine learning made simple. Use your preferred distributed ML framework like a lazy engineer. Getting Started • Highlight

Machine Learning Tooling 44 Nov 27, 2022
A Tools that help Data Scientists and ML engineers train and deploy ML models.

Domino Research This repo contains projects under active development by the Domino R&D team. We build tools that help Data Scientists and ML engineers

Domino Data Lab 73 Oct 17, 2022