Dataset Viewer

The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider removing the loading script and relying on automated data support (you can use convert_to_parquet from the datasets library). If this is not possible, please open a discussion for direct help.

Dataset Card for SmallNORB

Dataset Description

The SmallNORB dataset is a real-world stereo image dataset designed for benchmarking algorithms in disentangled representation learning and unsupervised representation learning. It was introduced by LeCun et al. (2004) for evaluating generic object recognition with invariance to pose and lighting.

Unlike synthetic datasets such as dSprites or MPI3D, which are generated as a complete Cartesian product of factors (i.e. every possible combination is present), SmallNORB consists of real photographs of physical toy objects under controlled variations, but not every combination of factors is present — for example, object instances are sampled randomly and the views (azimuth, elevation, lighting) do not form an exact grid.

Each sample contains two views:

  • Left image (96x96 grayscale)
  • Right image (96x96 grayscale)

Each image pair is associated with 4 known factors of variation and instance index:

  • category (object type)
  • instance (specific object instance)
  • elevation (camera tilt angle)
  • azimuth (camera rotation angle)
  • lighting (lighting condition)

The dataset allows researchers to evaluate representation learning on real-world 3D objects, under complex lighting and pose variations. SmallNORB provides an official train/test split. Typically, instance is not considered as a factor. Dataset visualization

Dataset Source

  • Homepage: https://cs.nyu.edu/~ylclab/data/norb-v1.0-small/
  • License: other. Small NORB is public domain, for research use.
  • Paper: Yann LeCun et al. Learning methods for generic object recognition with invariance to pose and lighting. CVPR 2004.

Dataset Structure

Factors Possible Classes (Indices) Values
category 0,...,4 airplane=0, car=1, truck=2, human=3, animal=4
instance 0,...,9 specific instance of object
elevation 0,...,8 9 elevation angles
azimuth 0,...,17 azimuth originally 0,2,...,34 → scaled to 0-17
lighting 0,...,5 6 lighting conditions

Note: The dataset is not a complete Cartesian product — instances and views are sampled in the original design. Each sample contains a left image and a right image, both corresponding to the same factors.

Example Usage

Below is a quick example of how to load this dataset via the Hugging Face Datasets library:

from datasets import load_dataset

# Load train set
train_ds = load_dataset("randall-lab/small-norb", split="train", trust_remote_code=True)

# Load test set
# test_ds = load_dataset("randall-lab/small-norb", split="test", trust_remote_code=True)

# Access a sample
example = train_ds[0]
left_image = example["left_image"]
right_image = example["right_image"]
label = example["label"]  # [category, elevation, azimuth, lighting]

# Label breakdown
category = example["category"]      # 0-4
instance = example["instance"]      # 0-9
elevation = example["elevation"]    # 0-8
azimuth = example["azimuth"]        # 0-17
lighting = example["lighting"]      # 0-5

# Visualize
left_image.show()
right_image.show()

print(f"Label (factors): {label}")

If you are using colab, you should update datasets to avoid errors

pip install -U datasets

Citation

@inproceedings{lecun2004learning,
  title={Learning methods for generic object recognition with invariance to pose and lighting},
  author={LeCun, Yann and Huang, Fu Jie and Bottou, Leon},
  booktitle={Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004.},
  volume={2},
  pages={II--104},
  year={2004},
  organization={IEEE}
}
Downloads last month
5