This repository contains an ongoing development of the next version of CPORT, and as such it is not stable yet. The codebase might change drastically in the future.
For the current stable version, please access the web service at https://alcazar.science.uu.nl/services/CPORT.
Please also refer to the original publication:
- de Vries, S. J. & Bonvin, A. M. J. J. CPORT: A Consensus Interface Predictor and Its Performance in Prediction-Driven Docking with HADDOCK. PLoS ONE vol. 6 e17695 (2011).
Server | Status | Implemented |
---|---|---|
WHISCY | 🟢 | ✔️ |
SCRIBER | 🟢 | ✔️ |
ISPRED4 | 🟢 | ✔️ |
SPPIDER | 🟢 | ✔️ |
meta-PPISP | 🟢 | ✔️ |
PredUs2 | 🟠 | |
Cons-PPISP | 🟢 | ✔️ |
PredictProtein | 🟢 | ✔️ |
PSIVER | 🔴 | ✔️ |
CSM-Potential | 🔴 | ✔️ |
ScanNet | 🟢 | ✔️ |
Please refer to INSTALL.md for installation instructions.
cport path/to/file/1PPE.pdb E
See all related data at https://github.com/haddocking/cport-data
pending
The cport_ispred4_scannet_sppider
model was developed to obtain the interface residues of proteins for protein docking, using the predictions of several webservers and a deep neural network. This model uses SCRIBER, SCANNET, ISPRED4, and SPPIDER.
The training and validation data can be found in "cport/data". Scripts for data preparetion step can be found in "cport/src/data_prep".
prepare_training_data.py: This script is used for combining the retrieved predictions of the receptors in BM5 and ARCTIC3D interface residues, and create a master ".csv" file for training the machine learning model.
prepare_validation_data.py: This script is used for combining the retrieved predictions of the Alphafold reference structures of the ligands in the BM5 and ARCTIC3D interface residues, and create a master ".csv" file for validating the machine learning model.
filter_pdb.py: This script is used on the AlphaFold validation data to distinguish between the residues according to their b-factor values. The residues with a b-factor value higher than 0.8 and below 0.8 are seperated and used for creating 2 different master ".csv" files. This is performed because the residues with lower b-factor values were not reliable for validation.
Scripts related to the machine learning model can be found in "cport/src/ml_models".
ml_dataprep.py: This script contains the codes for utilities used in building the model such as handling imbalance in the data, splitting the data into train-test data, etc. The default model is built and trained without any sampling on the data, with the ratio of the labels 0 and 1 being 2/1. If desired, the sampler can be used, created in lines 32-33.
tf_class.py: This is the script where the model is built and trained with the proper data. The model is built with certain parameters to maximize the efficiency of the predictions, considering the amount of the features and the training data. These parameters can be adjusted and a new model can be created by modifiying the following lines as desired:
line 44 -> modify units=16
line 47 -> modify loop variable to create more layers
line 48 -> modify units=16
line 55 -> modify learning_rate=1e-2
After the model has been built and trained, the model and the analysis graphs are saved into the path given in line 92. The training data for the model can be changed by changing the "pred_path" parameter in line 68. The sampling method can be modified by changing the "sampler" parameter in line 69
alphafold_data_prediction.py: This script is used to analyze the created model using the validation data of AlphaFold proteins.
Please have a look at CONTRIBUTE
If you encounter a bug or would like to request a feature, please open an ISSUE or a PR.