Create large-scale ML-driven multiscale simulation ensembles to study the interactions

Overview

MuMMI RAS v0.1

Released: Nov 16, 2021

MuMMI RAS is the application component of the MuMMI framework developed to create large-scale ML-driven multiscale simulation ensembles to study the interactions of RAS proteins and RAS-RAF protein complexes with lipid plasma membranes.

MuMMI framework was developed as part of the Pilot2 project of the Joint Design of Advanced Computing Solutions for Cancer funded jointly by the Department of Energy (DOE) and the National Cancer Institute (NCI).

The Pilot 2 project focuses on developing multiscale simulation models for understanding the interactions of the lipid plasma membrane with the RAS and RAF proteins. The broad computational tool development aims of this pilot are:

  • Developing scalable multi-scale molecular dynamics code that will automatically switch between phase field, coarse-grained and all-atom simulations.
  • Developing scalable machine learning and predictive models of molecular simulations to:
    • identify and quantify states from simulations
    • identify events from simulations that can automatically signal change of resolution between phase field, coarse-grained and all-atom simulations
    • aggregate information from the multi-resolution simulations to efficiently feedback to/from machine learning tools
  • Integrate sparse information from experiments with simulation data

MuMMI RAS defines the specific functionalities needed for the various components and scales of a target multiscale simulation. The application components need to define the scales, how to read the corresponding data, how to perform ML-based selection, how to run the simulations, how to perform analysis, and how to perform feedback. This code uses several utilities made available through "MuMMI Core".

Publications

MuMMI framework is described in the following publications.

  1. Bhatia et al. Generalizable Coordination of Large Multiscale Ensembles: Challenges and Learnings at Scale. In Proceedings of the ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, SC '21, Article No. 10, November 2021. doi:10.1145/3458817.3476210.

  2. Di Natale et al. A Massively Parallel Infrastructure for Adaptive Multiscale Simulations: Modeling RAS Initiation Pathway for Cancer. In Proceedings of the ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, SC '19, Article No. 57, November 2019. doi:10.1145/3295500.3356197.
    Best Paper at SC 2019.

  3. Ingólfsson et al. Machine Learning-driven Multiscale Modeling Reveals Lipid-Dependent Dynamics of RAS Signaling Protein. Proceedings of the National Academy of Sciences (PNAS), accepted, 2021. preprint.

  4. Reciprocal Coupling of Coarse-Grained and All-Atom scales. In preparation.

Installation

git clone https://github.com/mummi-framework/mummi-ras
cd mummi-ras
pip3 install .

export MUMMI_ROOT=/path/to/outputs
export MUMMI_CORE=/path/to/core/repo
export MUMMI_APP=/path/to/app/repo
export MUMMI_RESOURCES=/path/to/resources
The installaton process as described above installs the MuMMI framework. The simulation codes (gridsim2d, ddcMD, AMBER, GROMACS) are not included and are to be installed separately.
Spack installation. We are also working towards releasing the option of installing MuMMI and its dependencies through spack.

Authors and Acknowledgements

MuMMI was developed at Lawrence Livermore National Laboratory, in collaboration with Los Alamos National Laboratory, Oak Ridge National Laboratory, and International Business Machines. A list of main contributors is given below.

  • LLNL: Harsh Bhatia, Francesco Di Natale, Helgi I Ingólfsson, Joseph Y Moon, Xiaohua Zhang, Joseph R Chavez, Fikret Aydin, Tomas Oppelstrup, Timothy S Carpenter, Shiv Sundaram (previously LLNL), Gautham Dharuman (previously LLNL), Dong H Ahn, Stephen Herbein, Tom Scogland, Peer-Timo Bremer, and James N Glosli.

  • LANL: Chris Neale and Cesar Lopez

  • ORNL: Chris Stanley

  • IBM: Sara K Schumacher

MuMMI was funded by the Pilot2 project led by Dr. Fred Streitz (DOE) and Dr. Dwight Nissley (NIH). We acknowledge contributions from the entire Pilot 2 team.

This work was performed under the auspices of the U.S. Department of Energy (DOE) by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344, Los Alamos National Laboratory (LANL) under Contract DE-AC5206NA25396, and Oak Ridge National Laboratory under Contract DE-AC05-00OR22725.

Contact: Lawrence Livermore National Laboratory, 7000 East Avenue, Livermore, CA 94550.

Contributing

Contributions may be made through pull requests and/or issues on github.

License

MuMMI RAS is distributed under the terms of the MIT License.

Livermore Release Number: LLNL-CODE-827655

Comments
  • Are the trajectories in your publications publicly available?

    Are the trajectories in your publications publicly available?

    Hi, Congrats on the success, and huge thanks for making it open source. I wonder whether the trajectories in your publications are publicly available. Or are there any demo trajectories?

    I am a Ph.D. student at KAUST, using computer graphics to build and visualize mesoscale biology models, such as SARS-CoV-2 and bacteriophage T4. If possible, I (and my colleagues) would like to perform (multiscale, multi-representation, multi-granularity) visualization research on the trajectories you generated.

    Many thanks, Roden

    opened by RodenLuo 2
  • `flux` vs `slurm`

    `flux` vs `slurm`

    Hi,

    As flux is mentioned in the dependencies, is it possible to reproduce MuMMI RAS on a cluster that only has slurm?

    Workflow dependencies (e.g., python, flux, dynim, keras, etc.)

    Quoted from: https://github.com/mummi-framework/mummi-ras/blob/main/INSTALL.md

    Many thanks, Roden

    opened by RodenLuo 0
  • gridsim2d availability

    gridsim2d availability

    Hi, I wonder if the following code is available or not.

    gridsim2d: to be released shortly

    Quoted from: https://github.com/mummi-framework/mummi-ras/blob/main/INSTALL.md

    Thanks, Roden

    opened by RodenLuo 0
  • Patch for gromacs availability

    Patch for gromacs availability

    Hi, I wonder if the following patch is available or not.

    Note that we have a patch for gromacs installation for customization. To be open-sourced soon.

    Quoted from: https://github.com/mummi-framework/mummi-ras/blob/main/INSTALL.md

    Thanks, Roden

    opened by RodenLuo 0
  • Small scale test data for local deployment

    Small scale test data for local deployment

    Hi, I'm interested in deploying MuMMI on the KAUST IBEX cluster. It is mentioned in the installation doc that there is a small set of test data. Is it now publicly available? If not, is it possible for me to somehow access it so that I can perform a test run?

    Many thanks, Roden

    Again on lassen and on summit, we have created a small set of test data, which can be used to launch MuMMI at small scales. This (and the larger dataset) will be made public through NCI website. Until then, we can make this data available upon request.

    opened by RodenLuo 1
Releases(v1.0.0)
BigDL: Distributed Deep Learning Framework for Apache Spark

BigDL: Distributed Deep Learning on Apache Spark What is BigDL? BigDL is a distributed deep learning library for Apache Spark; with BigDL, users can w

4.1k Jan 09, 2023
A simple python program which predicts the success of a movie based on it's type, actor, actress and director

Movie-Success-Prediction A simple python program which predicts the success of a movie based on it's type, actor, actress and director. The program us

Mahalinga Prasad R N 1 Dec 17, 2021
Bodywork deploys machine learning projects developed in Python, to Kubernetes.

Bodywork deploys machine learning projects developed in Python, to Kubernetes. It helps you to: serve models as microservices execute batch jobs run r

Bodywork Machine Learning 409 Jan 01, 2023
A Python toolbox to churn out organic alkalinity calculations with minimal brain engagement.

Organic Alkalinity Sausage Machine A Python toolbox to churn out organic alkalinity calculations with minimal brain engagement. Getting started To mak

Charles Turner 1 Feb 01, 2022
Decision Weights in Prospect Theory

Decision Weights in Prospect Theory It's clear that humans are irrational, but how irrational are they? After some research into behavourial economics

Cameron Davidson-Pilon 32 Nov 08, 2021
Using Logistic Regression and classifiers of the dataset to produce an accurate recall, f-1 and precision score

Using Logistic Regression and classifiers of the dataset to produce an accurate recall, f-1 and precision score

Thines Kumar 1 Jan 31, 2022
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow

eXtreme Gradient Boosting Community | Documentation | Resources | Contributors | Release Notes XGBoost is an optimized distributed gradient boosting l

Distributed (Deep) Machine Learning Community 23.6k Jan 03, 2023
InfiniteBoost: building infinite ensembles with gradient descent

InfiniteBoost Code for a paper InfiniteBoost: building infinite ensembles with gradient descent (arXiv:1706.01109). A. Rogozhnikov, T. Likhomanenko De

Alex Rogozhnikov 183 Jan 03, 2023
hgboost - Hyperoptimized Gradient Boosting

hgboost is short for Hyperoptimized Gradient Boosting and is a python package for hyperparameter optimization for xgboost, catboost and lightboost using cross-validation, and evaluating the results o

Erdogan Taskesen 34 Jan 03, 2023
Gaussian Process Optimization using GPy

End of maintenance for GPyOpt Dear GPyOpt community! We would like to acknowledge the obvious. The core team of GPyOpt has moved on, and over the past

Sheffield Machine Learning Software 847 Dec 19, 2022
This is the material used in my free Persian course: Machine Learning with Python

This is the material used in my free Persian course: Machine Learning with Python

Yara Mohamadi 4 Aug 07, 2022
MBTR is a python package for multivariate boosted tree regressors trained in parameter space.

MBTR is a python package for multivariate boosted tree regressors trained in parameter space.

SUPSI-DACD-ISAAC 61 Dec 19, 2022
100 Days of Machine and Deep Learning Code

💯 Days of Machine Learning and Deep Learning Code MACHINE LEARNING TOPICS COVERED - FROM SCRATCH Linear Regression Logistic Regression K Means Cluste

Tanishq Gautam 66 Nov 02, 2022
Machine Learning Study 혼자 해보기

Machine Learning Study 혼자 해보기 기여자 (Contributors) ✨ Teddy Lee 🏠 HongJaeKwon 🏠 Seungwoo Han 🏠 Tae Heon Kim 🏠 Steve Kwon 🏠 SW Song 🏠 K1A2 🏠 Wooil

Teddy Lee 1.7k Jan 01, 2023
CyLP is a Python interface to COIN-OR’s Linear and mixed-integer program solvers (CLP, CBC, and CGL)

CyLP CyLP is a Python interface to COIN-OR’s Linear and mixed-integer program solvers (CLP, CBC, and CGL). CyLP’s unique feature is that you can use i

COIN-OR Foundation 161 Dec 14, 2022
Convoys is a simple library that fits a few statistical model useful for modeling time-lagged conversions.

Convoys is a simple library that fits a few statistical model useful for modeling time-lagged conversions. There is a lot more info if you head over to the documentation. You can also take a look at

Better 240 Dec 26, 2022
Software Engineer Salary Prediction

Based on 2021 stack overflow data, this machine learning web application helps one predict the salary based on years of experience, level of education and the country they work in.

Jhanvi Mimani 1 Jan 08, 2022
vortex particles for simulating smoke in 2d

vortex-particles-method-2d vortex particles for simulating smoke in 2d -vortexparticles_s

12 Aug 23, 2022
Flightfare-Prediction - It is a Flightfare Prediction Web Application Using Machine learning,Python and flask

Flight_fare-Prediction It is a Flight_fare Prediction Web Application Using Machine learning,Python and flask Using Machine leaning i have created a F

1 Dec 06, 2022
MegFlow - Efficient ML solutions for long-tailed demands.

Efficient ML solutions for long-tailed demands.

旷视天元 MegEngine 371 Dec 21, 2022