Skip to content
forked from GT4SD/gt4sd-core

GT4SD, an open-source library to accelerate hypothesis generation in the scientific discovery process.

License

Notifications You must be signed in to change notification settings

ACDBio/gt4sd-core

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GT4SD (Generative Toolkit for Scientific Discovery)

PyPI version Actions tests License: MIT Code style: black Contributions Docs Total downloads Monthly downloads Binder DOI 2022 IEEE Open Software Services Award

logo

The GT4SD (Generative Toolkit for Scientific Discovery) is an open-source platform to accelerate hypothesis generation in the scientific discovery process. It provides a library for making state-of-the-art generative AI models easier to use.

For full details on the library API and examples see the docs.

Installation

requirements

Currently gt4sd relies on:

  • python>=3.7,<3.9
  • pip>=19.1,<20.3

We are actively working on relaxing these, so stay tuned or help us with this by contributing to the project.

conda

The recommended way to install the gt4sd is to create a dedicated conda environment, this will ensure all requirements are satisfied. For CPU:

git clone https://github.com/GT4SD/gt4sd-core.git
cd gt4sd-core/
conda env create -f conda.yml
conda activate gt4sd

Note: by default gt4sd is installed with CPU requirements on linux systems. If you have GPU available, run:

git clone https://github.com/GT4SD/gt4sd-core.git
cd gt4sd-core/
conda env create -f conda_gpu.yml
conda activate gt4sd

And install the package via pip from PyPI:

pip install gt4sd

NOTE: In case you want to reuse an existing compatible environment (see requirements), you can use pip, but as of now (:eyes: on issue for changes), some dependencies require installation from GitHub, so for a complete setup install them with:

pip install -r vcs_requirements.txt

Development setup & installation

If you would like to contribute to the package, we recommend the following development setup:

conda env create -f conda.yml
conda activate gt4sd
# install gt4sd in editable mode
pip install --no-deps -e .

Learn more in CONTRIBUTING.md

Getting started

After install you can use gt4sd right away in your discovery workflows.

logo

Running inference pipelines in your python code

Running an algorithm is as easy as typing:

from gt4sd.algorithms.conditional_generation.paccmann_rl.core import (
    PaccMannRLProteinBasedGenerator, PaccMannRL
)
target = 'MVLSPADKTNVKAAWGKVGAHAGEYGAEALERMFLSFPTT'
# algorithm configuration with default parameters
configuration = PaccMannRLProteinBasedGenerator()
# instantiate the algorithm for sampling
algorithm = PaccMannRL(configuration=configuration, target=target)
items = list(algorithm.sample(10))
print(items)

Or you can use the ApplicationRegistry to run an algorithm instance using a serialized representation of the algorithm:

from gt4sd.algorithms.registry import ApplicationsRegistry
target = 'MVLSPADKTNVKAAWGKVGAHAGEYGAEALERMFLSFPTT'
algorithm = ApplicationsRegistry.get_application_instance(
    target=target,
    algorithm_type='conditional_generation',
    domain='materials',
    algorithm_name='PaccMannRL',
    algorithm_application='PaccMannRLProteinBasedGenerator',
    generated_length=32,
    # include additional configuration parameters as **kwargs
)
items = list(algorithm.sample(10))
print(items)

Running inference pipelines via the CLI command

GT4SD can run inference pipelines based on the gt4sd-inference CLI command. It allows to run all inference algorithms directly from the command line. To see which algorithms are available and how to use the CLI for your favorite model, check out examples/cli/README.md.

You can run inference pipelines simply typing:

gt4sd-inference --algorithm_name PaccMannRL --algorithm_application PaccMannRLProteinBasedGenerator --target MVLSPADKTNVKAAWGKVGAHAGEYGAEALERMFLSFPTT --number_of_samples 10

The command supports multiple parameters to select an algorithm and configure it for inference:

$ gt4sd-inference --help
usage: gt4sd-inference [-h] [--algorithm_type ALGORITHM_TYPE]
                       [--domain DOMAIN] [--algorithm_name ALGORITHM_NAME]
                       [--algorithm_application ALGORITHM_APPLICATION]
                       [--algorithm_version ALGORITHM_VERSION]
                       [--target TARGET]
                       [--number_of_samples NUMBER_OF_SAMPLES]
                       [--configuration_file CONFIGURATION_FILE]
                       [--print_info [PRINT_INFO]]

optional arguments:
  -h, --help            show this help message and exit
  --algorithm_type ALGORITHM_TYPE
                        Inference algorithm type, supported types:
                        conditional_generation, controlled_sampling,
                        generation, prediction. (default: None)
  --domain DOMAIN       Domain of the inference algorithm, supported types:
                        materials, nlp. (default: None)
  --algorithm_name ALGORITHM_NAME
                        Inference algorithm name. (default: None)
  --algorithm_application ALGORITHM_APPLICATION
                        Inference algorithm application. (default: None)
  --algorithm_version ALGORITHM_VERSION
                        Inference algorithm version. (default: None)
  --target TARGET       Optional target for generation represented as a
                        string. Defaults to None, it can be also provided in
                        the configuration_file as an object, but the
                        commandline takes precendence. (default: None)
  --number_of_samples NUMBER_OF_SAMPLES
                        Number of generated samples, defaults to 5. (default:
                        5)
  --configuration_file CONFIGURATION_FILE
                        Configuration file for the inference pipeline in JSON
                        format. (default: None)
  --print_info [PRINT_INFO]
                        Print info for the selected algorithm, preventing
                        inference run. Defaults to False. (default: False)

You can use gt4sd-inference to directly get information on the configuration parameters for the selected algorithm:

gt4sd-inference --algorithm_name PaccMannRL --algorithm_application PaccMannRLProteinBasedGenerator --print_info
INFO:gt4sd.cli.inference:Selected algorithm: {'algorithm_type': 'conditional_generation', 'domain': 'materials', 'algorithm_name': 'PaccMannRL', 'algorithm_application': 'PaccMannRLProteinBasedGenerator', 'algorithm_version': 'v0'}
INFO:gt4sd.cli.inference:Selected algorithm support the following configuration parameters:
{
 "batch_size": {
  "description": "Batch size used for the generative model sampling.",
  "title": "Batch Size",
  "default": 32,
  "type": "integer",
  "optional": true
 },
 "temperature": {
  "description": "Temperature parameter for the softmax sampling in decoding.",
  "title": "Temperature",
  "default": 1.4,
  "type": "number",
  "optional": true
 },
 "generated_length": {
  "description": "Maximum length in tokens of the generated molcules (relates to the SMILES length).",
  "title": "Generated Length",
  "default": 100,
  "type": "integer",
  "optional": true
 }
}
Target information:
{
 "target": {
  "title": "Target protein sequence",
  "description": "AA sequence of the protein target to generate non-toxic ligands against.",
  "type": "string"
 }
}

Running training pipelines via the CLI command

GT4SD provides a trainer client based on the gt4sd-trainer CLI command.

The trainer currently supports the following training pipelines:

  • language-modeling-trainer: language modelling via HuggingFace transfomers and PyTorch Lightning.
  • paccmann-vae-trainer: PaccMann VAE models.
  • granular-trainer: multimodal compositional autoencoders supporting MLP, RNN and Transformer layers.
  • guacamol-lstm-trainer: GuacaMol LSTM models.
  • moses-organ-trainer: Moses Organ implementation.
  • moses-vae-trainer: Moses VAE models.
  • torchdrug-gcpn-trainer: TorchDrug Graph Convolutional Policy Network model.
  • torchdrug-graphaf-trainer: TorchDrug autoregressive GraphAF model.
  • diffusion-trainer: Diffusers model.
  • gflownet-trainer: GFlowNet model.
$ gt4sd-trainer --help
usage: gt4sd-trainer [-h] --training_pipeline_name TRAINING_PIPELINE_NAME
                     [--configuration_file CONFIGURATION_FILE]

optional arguments:
  -h, --help            show this help message and exit
  --training_pipeline_name TRAINING_PIPELINE_NAME
                        Training type of the converted model, supported types:
                        granular-trainer, language-modeling-trainer, paccmann-
                        vae-trainer. (default: None)
  --configuration_file CONFIGURATION_FILE
                        Configuration file for the trainining. It can be used
                        to completely by-pass pipeline specific arguments.
                        (default: None)

To launch a training you have two options.

You can either specify the training pipeline and the path of a configuration file that contains the needed training parameters:

gt4sd-trainer  --training_pipeline_name ${TRAINING_PIPELINE_NAME} --configuration_file ${CONFIGURATION_FILE}

Or you can provide directly the needed parameters as arguments:

gt4sd-trainer  --training_pipeline_name language-modeling-trainer --type mlm --model_name_or_path mlm --training_file /path/to/train_file.jsonl --validation_file /path/to/valid_file.jsonl

To get more info on a specific training pipeleins argument simply type:

gt4sd-trainer --training_pipeline_name ${TRAINING_PIPELINE_NAME} --help

Saving a trained algorithm for inference via the CLI command

Once a training pipeline has been run via the gt4sd-trainer, it's possible to save the trained algorithm via gt4sd-saving for usage in compatible inference pipelines.

Here a small example for PaccMannGP algorithm (paper).

You can train a model with gt4sd-trainer (quick training using few data, not really recommended for a realistic model ⚠️):

gt4sd-trainer  --training_pipeline_name paccmann-vae-trainer --epochs 250 --batch_size 4 --n_layers 1 --rnn_cell_size 16 --latent_dim 16 --train_smiles_filepath src/gt4sd/training_pipelines/tests/molecules.smi --test_smiles_filepath src/gt4sd/training_pipelines/tests/molecules.smi --model_path /tmp/gt4sd-paccmann-gp/ --training_name fast-example --eval_interval 15 --save_interval 15 --selfies

Save the model with the compatible inference pipeline using gt4sd-saving:

gt4sd-saving --training_pipeline_name paccmann-vae-trainer --model_path /tmp/gt4sd-paccmann-gp/ --training_name fast-example --target_version fast-example-v0 --algorithm_application PaccMannGPGenerator

Run the algorithm via gt4sd-inference (again the model produced in the example is trained on dummy data and will give dummy outputs, do not use it as is 🙅):

gt4sd-inference --algorithm_name PaccMannGP --algorithm_application PaccMannGPGenerator --algorithm_version fast-example-v0 --number_of_samples 5  --target '{"molwt": {"target": 60.0}}'

Uploading a trained algorithm on a public hub via the CLI command

You can upload trained and finetuned models easily in the public hub using gt4sd-upload. The syntax follows the saving pipeline:

gt4sd-upload --training_pipeline_name paccmann-vae-trainer --model_path /tmp/gt4sd-paccmann-gp --training_name fast-example --target_version fast-example-v0 --algorithm_application PaccMannGPGenerator

NOTE: GT4SD can be configured to upload models to a custom or self-hosted COS. An example on self-hosting locally a COS (minio) where to upload your models can be found here.

Computing properties

You can compute properties of your generated samples using the gt4sd.properties submodule:

>>>from gt4sd.properties import PropertyPredictorRegistry
>>>similarity_predictor = PropertyPredictorRegistry.get_property_predictor("similarity_seed", {"smiles" : "C1=CC(=CC(=C1)Br)CN"})
>>>similarity_predictor("CCO")
0.0333
>>># let's inspect what other parameters we can set for similarity measuring
>>>similarity_predictor = PropertyPredictorRegistry.get_property_predictor("similarity_seed", {"smiles" : "C1=CC(=CC(=C1)Br)CN", "fp_key": "ECFP6"})
>>>similarity_predictor("CCO")
>>># inspect parameters
>>>PropertyPredictorRegistry.get_property_predictor_parameters_schema("similarity_seed")
'{"title": "SimilaritySeedParameters", "description": "Abstract class for property computation.", "type": "object", "properties": {"smiles": {"title": "Smiles", "example": "c1ccccc1", "type": "string"}, "fp_key": {"title": "Fp Key", "default": "ECFP4", "type": "string"}}, "required": ["smiles"]}'
>>># predict other properties
>>>qed = PropertyPredictorRegistry.get_property_predictor("qed")
>>>qed('CCO')
0.4068
>>># list properties
>>>PropertyPredictorRegistry.list_available()
['activity_against_target',
 'aliphaticity',
 ...
 'scscore',
 'similarity_seed',
 'tpsa',
 'weight']

Additional examples

Find more examples in notebooks

You can play with them right away using the provided Dockerfile, simply build the image and run it to explore the examples using Jupyter:

docker build -f Dockerfile -t gt4sd-demo .
docker run -p 8888:8888 gt4sd-demo

Supported packages

Beyond implementing various generative modeling inference and training pipelines GT4SD is designed to provide a high-level API that implement an harmonized interface for several existing packages:

  • GuacaMol: inference pipelines for the baselines models and training pipelines for LSTM models.
  • Moses: inference pipelines for the baselines models and training pipelines for VAEs and Organ.
  • TorchDrug: inference and training pipelines for GCPN and GraphAF models. Training pipelines support custom datasets as well as datasets native in TorchDrug.
  • MoLeR: inference pipelines for MoLeR (MOlecule-LEvel Representation) generative models for de-novo and scaffold-based generation.
  • TAPE: encoder modules compatible with the protein language models.
  • PaccMann: inference pipelines for all algorithms of the PaccMann family as well as training pipelines for the generative VAEs.
  • transformers: training and inference pipelines for generative models from HuggingFace Models
  • diffusers: training and inference pipelines for generative models from Diffusers Models
  • GFlowNets: training and inference pipeline for Generative Flow Networks
  • MolGX: training and inference pipelines to generate small molecules satisfying target properties. The full implementation of MolGX, including additional functionalities, is available here.

References

If you use gt4sd in your projects, please consider citing the following:

@software{GT4SD,
author = {GT4SD Team},
month = {2},
title = {{GT4SD (Generative Toolkit for Scientific Discovery)}},
url = {https://github.com/GT4SD/gt4sd-core},
version = {main},
year = {2022}
}

@article{manica2022gt4sd,
  title={GT4SD: Generative Toolkit for Scientific Discovery},
  author={Manica, Matteo and Cadow, Joris and Christofidellis, Dimitrios and Dave, Ashish and Born, Jannis and Clarke, Dean and Teukam, Yves Gaetan Nana and Hoffman, Samuel C and Buchan, Matthew and Chenthamarakshan, Vijil and others},
  journal={arXiv preprint arXiv:2207.03928},
  year={2022}
}

License

The gt4sd codebase is under MIT license. For individual model usage, please refer to the model licenses found in the original packages.

About

GT4SD, an open-source library to accelerate hypothesis generation in the scientific discovery process.

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Jupyter Notebook 65.2%
  • Python 34.8%