--- license: apache-2.0 --- # Dataset Card for 3dshapes ## Dataset Description The **3dshapes dataset** is a **synthetic 3D object image dataset** designed for benchmarking algorithms in **disentangled representation learning** and **unsupervised representation learning**. It was introduced in the **FactorVAE** paper [[Kim & Mnih, ICML 2018](https://proceedings.mlr.press/v80/kim18b.html)], as one of the standard testbeds for learning interpretable and disentangled latent factors. The dataset consists of images of **3D procedurally generated scenes**, where 6 **ground-truth independent factors of variation** are explicitly controlled: - **Floor color** (hue) - **Wall color** (hue) - **Object color** (hue) - **Object size** (scale) - **Object shape** (categorical) - **Object orientation** (rotation angle) **3dshapes is generated as a full Cartesian product of all factor combinations**, making it perfectly suited for systematic evaluation of disentanglement. The dataset contains **480,000 images** at a resolution of **64×64 pixels**, covering **all possible combinations of the 6 factors exactly once**. The images are stored in **row-major order** according to the factor sweep, enabling precise control over factor-based evaluation. ![Dataset Visualization](https://huggingface.co/datasets/randall-lab/shapes3d/resolve/main/3dshapes.gif) ## Dataset Source - **Homepage**: [https://github.com/deepmind/3dshapes-dataset](https://github.com/deepmind/3dshapes-dataset) - **License**: [Apache License 2.0](http://www.apache.org/licenses/LICENSE-2.0) - **Paper**: Hyunjik Kim & Andriy Mnih. _Disentangling by Factorising_. ICML 2018. ## Dataset Structure |Factors|Possible Values| |---|---| |floor_color (hue)| 10 values linearly spaced in [0, 1] | |wall_color (hue)| 10 values linearly spaced in [0, 1] | |object_color (hue)| 10 values linearly spaced in [0, 1] | |scale| 8 values linearly spaced in [0.75, 1.25] | |shape| 4 values: 0, 1, 2, 3 | |orientation| 15 values linearly spaced in [-30, 30] | Each image corresponds to a unique combination of these **6 factors**. The images are stored in a **row-major order** (fastest-changing factor is `orientation`, slowest-changing factor is `floor_color`). ### Why no train/test split? The 3dshapes dataset does not provide an official train/test split. It is designed for **representation learning research**, where the goal is to learn disentangled and interpretable latent factors. Since the dataset is a **complete Cartesian product of all factor combinations**, models typically require access to the full dataset to explore factor-wise variations. ## Example Usage Below is a quick example of how to load this dataset via the Hugging Face Datasets library: ```python from datasets import load_dataset # Load the dataset dataset = load_dataset("randall-lab/shapes3d", split="train", trust_remote_code=True) # Access a sample from the dataset example = dataset[0] image = example["image"] label = example["label"] # Value labels: [floor_hue, wall_hue, object_hue, scale, shape, orientation] label_index = example["label_index"] # Index labels: [floor_idx, wall_idx, object_idx, scale_idx, shape_idx, orientation_idx] # Label Value floor_value = example["floor"] # 0-1 wall_value = example["wall"] # 0-1 object_value = example["object"] # 0-1 scale_value = example["scale"] # 0.75-1.25 shape_value = example["shape"] # 0,1,2,3 orientation_value = example["orientation"] # -30 - 30 # Label index floor_idx = example["floor_idx"] # 0-9 wall_idx = example["wall_idx"] # 0-9 object_idx = example["object_idx"] # 0-9 scale_idx = example["scale_idx"] # 0-7 shape_idx = example["shape_idx"] # 0-3 orientation_idx = example["orientation_idx"] # 0-14 image.show() # Display the image print(f"Label (factor values): {label}") print(f"Label (factor indices): {label_index}") ``` If you are using colab, you should update datasets to avoid errors ``` pip install -U datasets ``` ## Citation ``` @InProceedings{pmlr-v80-kim18b, title = {Disentangling by Factorising}, author = {Kim, Hyunjik and Mnih, Andriy}, booktitle = {Proceedings of the 35th International Conference on Machine Learning}, pages = {2649--2658}, year = {2018}, editor = {Dy, Jennifer and Krause, Andreas}, volume = {80}, series = {Proceedings of Machine Learning Research}, month = {10--15 Jul}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v80/kim18b/kim18b.pdf}, url = {https://proceedings.mlr.press/v80/kim18b.html}, abstract = {We define and address the problem of unsupervised learning of disentangled representations on data generated from independent factors of variation. We propose FactorVAE, a method that disentangles by encouraging the distribution of representations to be factorial and hence independent across the dimensions. We show that it improves upon beta-VAE by providing a better trade-off between disentanglement and reconstruction quality and being more robust to the number of training iterations. Moreover, we highlight the problems of a commonly used disentanglement metric and introduce a new metric that does not suffer from them.} } ```