This repository provides a comprehensive suite of tools for protein design evaluation, integrating multiple state-of-the-art models with standardized pipelines. It supports both monomer and binder design, enabling thorough assessment across diverse aspects of protein design.
The codebase is organized into three main components:
-
metrics
: Scripts for evaluating multiple aspects of protein design, including sequence quality, structure quality, and designability. -
tasks
: Pipelines for executing specific protein design evaluations (e.g., monomer, binder). -
tools
: Wrappers for external models (e.g., Protenix, ProteinMPNN, AlphaFold2, ESMFold) to streamline integration.
Task | Sequence Generation | Structure Consistency |
---|---|---|
Monomer | ProteinMPNN | πΉ ESMFold |
Binder | ProteinMPNN | πΉ AlphaFold2 πΉ Protenix |
- Install Protenix (version >= 0.5.3) β a high-accuracy structure prediction tool:
pip install git+https://github.com/bytedance/Protenix.git@dev_design
Note: You need to install Protenix from the dev_design
branch. If you encounter any issues during installation, please refer to the official Protenix Installation Guide for detailed instructions.
- Install additional dependencies for ProteinMPNN, AlphaFold2, and ESMFold:
pip install git+https://github.com/sokrypton/ColabDesign.git --no-deps # tested with colabdesign-1.1.3
pip install "jax[cuda]" -f https://storage.googleapis.com/jax-releases/jax_cuda_releases.html
pip install transformers==4.51.3
- Download model weights:
The default weight paths are defined in pxdbench/globals.py
.
After downloading, update the variables in globals.py
to reflect your local directory structure.
Model weights for external tools are expected to be organized in a directory as follows:
βββ af2
β βββ LICENSE
β βββ params_model_{1..5}.npz
β βββ params_model_{1..5}_ptm.npz
β βββ params_model_{1..5}_multimer_v3.npz
β
βββ esmfold
β βββ config.json
β βββ pytorch_model.bin
β βββ special_tokens_map.json
β βββ tokenizer_config.json
β βββ vocab.txt
β
βββ mpnn
β βββ ca_model_weights/...
β βββ soluble_model_weights/...
β βββ vanilla_model_weights/...
Note: Required Protenix files (weights, CCD files, etc.) will be auto-downloaded on the first evaluation run.
We provide demo scripts for both monomer and binder design evaluation. Monomer evaluation example:
bash monomer_eval_demo.sh
Binder evaluation example:
bash binder_eval_demo.sh
PXDesignBench supports multiple input modes, allowing you to evaluate protein designs flexibly.
The basic CLI arguments are:
--data_dir
: Directory containing input structures.--dump_dir
: Output directory for evaluation results.--is_mmcif
: Flag indicating whether input files are in mmCIF format (otherwise assumed PDB).
JSON-based Input One can also provide a JSON configuration file describing the evaluation task. This format allows fine-grained control over task parameters and is particularly useful for batch evaluation.
Example JSON:
{
"task": ..., # "monomer" or "binder"
"pdb_dir": ..., # directory containing the input PDB structures
"name": ..., # name of the task; used to locate "{pdb_dir}/{pdb_name}.pdb"
"pdb_names": ..., # list of PDB file names to evaluate
"cond_chains": ..., # list of condition chains (only for binder evaluation)
"binder_chains": ..., # list of binder chains (binder evaluation only; currently supports **one** binder chain)
"out_dir": ... # directory to store evaluation results
}
Key points:
- Binder tasks require
binder_chains
to be explicitly specified. - Currently only one binder chain is supported; all other chains will be treated as condition chains.
pdb_names
defines the exact structures to evaluate. If omitted, all files inpdb_dir
with valid suffixes will be evaluated.
Directory-based Input
Instead of JSON, one may provide a directory path directly to --data_dir
.
In this case:
- If
file_name_list
is provided, only matching files will be evaluated. - Otherwise, all files in the directory with valid extensions will be included.
Binder evaluation supports passing a JSON file to specify additional metadata beyond the default inputs.
This is useful for advanced scenarios such as:
-
Evaluating cropped sequences
- If the sequence to be evaluated is a cropped segment of an original sequence, you can provide the full original sequence along with a
crop
field to specify the range used in evaluation. - The crop range can be multiple ranges such as
"1-120,130-150"
(comma-separated ranges, 1-based indexing, inclusive).
- If the sequence to be evaluated is a cropped segment of an original sequence, you can provide the full original sequence along with a
-
Providing precomputed MSA for Protenix filter
- The Protenix filter requires the target chain's MSA.
- By default, the evaluation script will automatically call the Protenix MSA server to compute the MSA.
- If you have already computed the MSA locally, you can skip the server call by specifying the
msa
field with:precomputed_msa_dir
: Path to the local MSA directory.pairing_db
:uniref100
.
Example JSON input:
[
{
"proteinChain": {
"sequence": "NAFTVTVPKDLYVVEYGSNMTIECKFPVEKQLDLAALIVYWEMEDKNIIQFVHGEEDLKVQHSSYRQRARLLKDQLSLGNAALQITDVKLQDAGVYRCMISYGGADYKRITVKVNA",
"label_asym_id": ["A0"],
"use_msa": true,
"msa": {
"precomputed_msa_dir": "examples/msa/PDL1/0",
"pairing_db": "uniref100"
},
"crop": "1-116"
}
}
]
PXDesignBench exposes device IDs for each integrated model, enabling:
- Deployment across multiple GPUs for parallel evaluation.
- Integration into DDP (Distributed Data Parallel) training pipelines for online evaluation tracking.
For example, the following is a pseudocode snippet illustrating online evaluation tracking in the DDP model training pipeline:
from pxdbench.run import run_task
from protenix.utils.distributed import DIST_WRAPPER
@torch.no_grad()
def _evaluate_design(self):
# Sampling stage
self._inference_design()
DIST_WRAPPER.barrier()
# Load task list
with open(self._get_eval_json_path(), "r") as f:
all_eval_tasks = json.load(f)
# Distribute tasks among workers
task_indices = list(range(len(all_eval_tasks)))[DIST_WRAPPER.rank::DIST_WRAPPER.world_size]
if not task_indices:
self.local_print("No task to evaluate, skipping.")
results = []
else:
self.local_print(f"Running {len(task_indices)} tasks...")
results = [
run_task(all_eval_tasks[i], self.configs.eval, device_id=DIST_WRAPPER.local_rank)
for i in task_indices
]
# Gather results
all_eval_results = DIST_WRAPPER.all_gather_object(results)
# Custom logging...
- If
use_gt_seq=True
, the sequence from the input structure is used directly. - If
use_gt_seq=False
, the tool will first run the assigned sequence generation model (e.g., ProteinMPNN) to generate sequences. - Structure quality is then assessed using the corresponding structure prediction models:
- Monomer β ESMFold
- Binder β AlphaFold2 / Protenix
- Metrics include:
- Self-consistency between predicted structures
- Confidence scores from structure predictors (e.g., pLDDT, ipTM)
- Shape-based metrics, e.g. secondary structure content (Ξ±-helix, Ξ²-sheet, loop ratios), radius of gyration, etc.
- Results are stored in a summary CSV for downstream analysis.
PXDesignBench provides scripts for analyzing the diversity and novelty of generated protein structures.
To enable Foldseek-based diversity and novelty calculations, you must first install Foldseek, a structural alignment and similarity search tool.
Foldseek is not bundled with PXDesignBench and must be installed separately.
Please follow the official guide here: Foldseek Installation.
Examples:
- Monomer post-processing
python3 pxdbench/scripts/postprocess_monomer.py --input_dir examples/monomer
- Binder post-processing
python3 pxdbench/scripts/postprocess_binder.py --input_dir examples/binder --is_mmcif true
If you use this repository, please cite the following works:
PXDesign
@article{ren2025pxdesign,
title={PXDesign: Fast, Modular, and Accurate De Novo Design of Protein Binders},
author={Ren, Milong and Sun, Jinyuan and Guan, Jiaqi and Liu, Cong and Gong, Chengyue and Wang, Yuzhe and Wang, Lan and Cai, Qixu and Chen, Xinshi and Xiao, Wenzhi},
journal={bioRxiv},
pages={2025--08},
year={2025},
publisher={Cold Spring Harbor Laboratory}
}
Protenix
@article{bytedance2025protenix,
title={Protenix - Advancing Structure Prediction Through a Comprehensive AlphaFold3 Reproduction},
author={ByteDance AML AI4Science Team and Chen, Xinshi and Zhang, Yuxuan and Lu, Chan and Ma, Wenzhi and Guan, Jiaqi and Gong, Chengyue and Yang, Jincai and Zhang, Hanyu and Zhang, Ke and Wu, Shenghao and Zhou, Kuangqi and Yang, Yanping and Liu, Zhenyu and Wang, Lan and Shi, Bo and Shi, Shaochen and Xiao, Wenzhi},
year={2025},
journal={bioRxiv},
publisher={Cold Spring Harbor Laboratory},
doi={10.1101/2025.01.08.631967},
URL={https://www.biorxiv.org/content/early/2025/01/11/2025.01.08.631967},
elocation-id={2025.01.08.631967},
eprint={https://www.biorxiv.org/content/early/2025/01/11/2025.01.08.631967.full.pdf},
}
ProteinMPNN
@article{dauparas2022robust,
title={Robust deep learning--based protein sequence design using ProteinMPNN},
author={Dauparas, Justas and Anishchenko, Ivan and Bennett, Nathaniel and Bai, Hua and Ragotte, Robert J and Milles, Lukas F and Wicky, Basile IM and Courbet, Alexis and de Haas, Rob J and Bethel, Neville and others},
journal={Science},
volume={378},
number={6615},
pages={49--56},
year={2022},
publisher={American Association for the Advancement of Science}
}
ESMFold
@article{lin2023evolutionary,
title={Evolutionary-scale prediction of atomic-level protein structure with a language model},
author={Lin, Zeming and Akin, Halil and Rao, Roshan and Hie, Brian and Zhu, Zhongkai and Lu, Wenting and Smetanin, Nikita and Verkuil, Robert and Kabeli, Ori and Shmueli, Yaniv and others},
journal={Science},
volume={379},
number={6637},
pages={1123--1130},
year={2023},
publisher={American Association for the Advancement of Science}
}
AlphaFold2
@article{jumper2021highly,
title={Highly accurate protein structure prediction with AlphaFold},
author={Jumper, John and Evans, Richard and Pritzel, Alexander and Green, Tim and Figurnov, Michael and Ronneberger, Olaf and Tunyasuvunakool, Kathryn and Bates, Russ and {\v{Z}}{\'\i}dek, Augustin and Potapenko, Anna and others},
journal={nature},
volume={596},
number={7873},
pages={583--589},
year={2021},
publisher={Nature Publishing Group UK London}
}
We welcome contributions from the community to help improve the evaluation tool!
π Check out the Contributing Guide to get started.
β
Code Quality:
We use pre-commit
hooks to ensure consistency and code quality. Please install them before making commits:
pip install pre-commit
pre-commit install
We are committed to fostering a welcoming and inclusive environment. Please review our Code of Conduct for guidelines on how to participate respectfully.
If you discover a potential security issue in this project, or think you may have discovered a security issue, we ask that you notify Bytedance Security via our security center or vulnerability reporting email.
Please do not create a public GitHub issue.
This project is licensed under the Apache 2.0 License. It is free for both academic research and commercial use.