File size: 4,673 Bytes
163cef7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1771fcf
163cef7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
---
tags:
- protein-protein interaction
- binding affinity
- protein language models
- drug discovery
- bioinformatics
- multi-chain proteins
- ppb affinity
license: cc-by-nc-sa-4.0
pipeline_tag: regression
task_categories:
- text-classification
---

## Dataset Description

This repository provides several enhanced versions of the **PPB-Affinity dataset**, ready for both sequence and structure-based modeling of multi-chain protein-protein interactions. The original PPB-Affinity dataset was introduced in the paper "[PPB-Affinity: Protein-Protein Binding Affinity dataset for AI-based protein drug discovery](https://www.nature.com/articles/s41597-024-03997-4)".

This version of the dataset was prepared for the study: "*[Beyond Simple Concatenation: Fairly Assessing PLM Architectures for Multi-Chain Protein-Protein Interactions Prediction](https://arxiv.org/pdf/2505.20036)*."

The primary enhancements in this repository include:
* Various levels of data filtration and processing.
* The addition of pre-extracted "Ligand Sequences" and "Receptor Sequences" columns, making the dataset ready for use with sequence-based models without requiring PDB file parsing. For complexes with multiple ligand or receptor chains, the sequences are comma-separated.

## Dataset Configurations

This dataset offers four distinct configurations:

### 1. `raw`
* **Description**: Minimally processed data from the original PPB-Affinity dataset. Only annotation inconsistencies have been resolved (see Section 2.1.1 of "*Beyond Simple Concatenation...*" for details).
* **Size**: 12,048 entries.
* **Splits**: Contains a single `train` split encompassing all entries.
* **How to load**:
    ```python
    from datasets import load_dataset

    raw_ds = load_dataset(
        "proteinea/ppb_affinity",
        name="raw",
        trust_remote_code=True
    )["train"]
    ```

### 2. `raw_rec`
* **Description**: Similar to the `raw` version, but with an additional step to recover missing residues in the protein sequences (see Section 2.1.2 of "*Beyond Simple Concatenation...*" for details).
* **Size**: 12,048 entries.
* **Splits**: Contains a single `train` split.
* **How to load**:
    ```python
    from datasets import load_dataset

    raw_rec_ds = load_dataset(
        "proteinea/ppb_affinity",
        name="raw_rec",
        trust_remote_code=True
    )["train"]
    ```

### 3. `filtered`
* **Description**: This version includes additional cleaning and filtration steps applied to the raw with missing residues recovered data (see Section 2.1.2 of "*Beyond Simple Concatenation...*" for details on filtration). It comes with pre-defined train, validation, and test splits (see Section 2.1.3 of "*Beyond Simple Concatenation...*" for splitting methodology).
* **Size**:
    * Train: 6,485 entries
    * Validation: 965 entries
    * Test: 757 entries
* **Splits**: `train`, `validation`, `test`.
* **How to load**:
    ```python
    from datasets import load_dataset

    dataset_dict = load_dataset(
        "proteinea/ppb_affinity",
        name="filtered",
        trust_remote_code=True
    )
    train_ds = dataset_dict["train"]
    val_ds = dataset_dict["validation"]
    test_ds = dataset_dict["test"]
    ```

### 4. `filtered_random`
* **Description**: This version uses the same cleaned and filtered entries as the `filtered` configuration but provides random 80%-10%-10% splits for train, validation, and test, respectively. The shuffling is performed with a fixed seed (42) for reproducibility.
* **Size**: Same total entries as `filtered`, split as:
    * Train: 6,565 entries
    * Validation: 820 entries
    * Test: 822 entries
* **Splits**: `train`, `validation`, `test`.
* **How to load**:
    ```python
    from datasets import load_dataset

    dataset_dict = load_dataset(
        "proteinea/ppb_affinity",
        name="filtered_random",
        trust_remote_code=True
    )
    train_ds = dataset_dict["train"]
    val_ds = dataset_dict["validation"]
    test_ds = dataset_dict["test"]
    ```

## Data Fields

All configurations share a common set of columns. These include columns from the original PPB-Affinity dataset (refer to the original paper for more details), plus two new sequence columns:

* **`Ligand Sequences`**: `string` - Comma-separated amino acid sequences of the ligand chain(s).
* **`Receptor Sequences`**: `string` - Comma-separated amino acid sequences of the receptor chain(s).

**Note on Sequences**: When multiple ligand or receptor chains are present in a complex, their respective amino acid sequences are concatenated with a comma (`,`) as a separator in the "Ligand Sequences" and "Receptor Sequences" fields.

---