The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: DatasetGenerationError
Exception: ArrowNotImplementedError
Message: Unsupported cast from double to halffloat using function cast_half_float
Traceback: Traceback (most recent call last):
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1766, in _prepare_split_single
writer.write(example, key)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 500, in write
self.write_examples_on_file()
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 458, in write_examples_on_file
self.write_batch(batch_examples=batch_examples)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 568, in write_batch
arrays.append(pa.array(typed_sequence))
File "pyarrow/array.pxi", line 247, in pyarrow.lib.array
File "pyarrow/array.pxi", line 112, in pyarrow.lib._handle_arrow_array_protocol
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 208, in __arrow_array__
out = cast_array_to_feature(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature
casted_array_values = _c(array.values, feature.feature)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature
casted_array_values = _c(array.values, feature.feature)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature
casted_array_values = _c(array.values, feature.feature)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2116, in cast_array_to_feature
return array_cast(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1963, in array_cast
return array.cast(pa_type)
File "pyarrow/array.pxi", line 996, in pyarrow.lib.Array.cast
File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/compute.py", line 404, in cast
return call_function("cast", [arr], options, memory_pool)
File "pyarrow/_compute.pyx", line 590, in pyarrow._compute.call_function
File "pyarrow/_compute.pyx", line 385, in pyarrow._compute.Function.call
File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
pyarrow.lib.ArrowNotImplementedError: Unsupported cast from double to halffloat using function cast_half_float
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1775, in _prepare_split_single
num_examples, num_bytes = writer.finalize()
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 599, in finalize
self.write_examples_on_file()
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 458, in write_examples_on_file
self.write_batch(batch_examples=batch_examples)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 568, in write_batch
arrays.append(pa.array(typed_sequence))
File "pyarrow/array.pxi", line 247, in pyarrow.lib.array
File "pyarrow/array.pxi", line 112, in pyarrow.lib._handle_arrow_array_protocol
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 208, in __arrow_array__
out = cast_array_to_feature(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature
casted_array_values = _c(array.values, feature.feature)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature
casted_array_values = _c(array.values, feature.feature)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2075, in cast_array_to_feature
casted_array_values = _c(array.values, feature.feature)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2116, in cast_array_to_feature
return array_cast(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1804, in wrapper
return func(array, *args, **kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 1963, in array_cast
return array.cast(pa_type)
File "pyarrow/array.pxi", line 996, in pyarrow.lib.Array.cast
File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/compute.py", line 404, in cast
return call_function("cast", [arr], options, memory_pool)
File "pyarrow/_compute.pyx", line 590, in pyarrow._compute.call_function
File "pyarrow/_compute.pyx", line 385, in pyarrow._compute.Function.call
File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
pyarrow.lib.ArrowNotImplementedError: Unsupported cast from double to halffloat using function cast_half_float
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1524, in compute_config_parquet_and_info_response
parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet(
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1099, in stream_convert_to_parquet
builder._prepare_split(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1627, in _prepare_split
for job_id, done, content in self._prepare_split_single(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1784, in _prepare_split_single
raise DatasetGenerationError("An error occurred while generating the dataset") from e
datasets.exceptions.DatasetGenerationError: An error occurred while generating the datasetNeed help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
__key__
string | __url__
string | img.png
image | latent.pth
sequence |
|---|---|---|---|
0
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[0.56640625,1.4072265625,0.38525390625,1.400390625,3.142578125,1.998046875,-1.7041015625,1.1953125(...TRUNCATED)
|
|
1
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[-17.703125,-19.6875,-16.140625,-20.765625,-14.4453125,-19.9375,-19.3125,-18.234375,-17.0625,-18.4(...TRUNCATED)
|
|
2
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[5.99609375,3.982421875,5.78125,3.6171875,6.734375,4.03125,1.73046875,5.59765625,3.7421875,2.70312(...TRUNCATED)
|
|
3
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[3.908203125,3.609375,4.52734375,0.8994140625,6.8125,4.09765625,0.56005859375,5.35546875,4.4765625(...TRUNCATED)
|
|
4
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[19.0,16.109375,17.140625,18.65625,15.2421875,22.703125,19.6875,18.375,17.25,21.8125,20.5625,15.89(...TRUNCATED)
|
|
5
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[-11.5859375,-13.578125,-15.2890625,-12.6484375,-16.9375,-12.7109375,-12.1875,-13.5390625,-11.1328(...TRUNCATED)
|
|
6
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[13.1953125,11.1484375,9.3515625,12.1328125,8.0390625,9.90625,11.6875,12.265625,12.5078125,10.25,1(...TRUNCATED)
|
|
7
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[-11.2890625,-14.796875,-15.390625,-6.09765625,10.8671875,4.55859375,0.137451171875,1.25,3.828125,(...TRUNCATED)
|
|
8
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[6.59375,6.22265625,-0.6640625,-1.6787109375,-7.44921875,-7.93359375,-3.994140625,-2.41796875,-3.7(...TRUNCATED)
|
|
9
|
hf://datasets/Birchlabs/sdxl-latents-ffhq@667d9927f3d7eb419d21b93e9aec9f3ffc9bb952/00000.tar
| [[[-1.9423828125,-1.6826171875,5.3671875,0.46044921875,6.59375,9.890625,4.09375,-0.3984375,4.1210937(...TRUNCATED)
|
https://github.com/NVlabs/ffhq-dataset samples encoded to float16 SDXL latents via Ollin VAE.
Dataset created using this script.
VAE encoder used NATTEN attention, kernel size 17.
Didn't bother saving mean & logvar, because variance is low enough it's not worth the doubling of filesize to retain.
Sampled from diagonal gaussian distribution, saved the resulting latents.
Also kept the original image.
Schema/usage:
from typing import TypedDict, Iterator
from webdataset import WebDataset
Sample = TypedDict('Sample', {
'__key__': str,
'__url__': str,
'img.png': bytes, # PIL image, serialized. 1024*1024px
'latent.pth': bytes, # FloatTensor, serialized. 128*128 latents
})
it: Iterator[Sample] = WebDataset('{00000..00035}.tar')
for sample in it:
pass
# avg/val.pt (mean):
[-2.8982300758361816, -0.9609659910202026, 0.2416578084230423, -0.307400107383728]
# avg/sq.pt:
[65.80902099609375, 32.772762298583984, 36.080204010009766, 25.072017669677734]
# std
# (sq - val**2)**.5
[7.5768914222717285, 5.643518924713135, 6.001816749572754, 4.997751712799072]
# 1/std
[0.13198024034500122, 0.17719440162181854, 0.16661621630191803, 0.2000899761915207]
Flickr-Faces-HQ Dataset (FFHQ)
Flickr-Faces-HQ (FFHQ) is a high-quality image dataset of human faces, originally created as a benchmark for generative adversarial networks (GAN):
A Style-Based Generator Architecture for Generative Adversarial Networks
Tero Karras (NVIDIA), Samuli Laine (NVIDIA), Timo Aila (NVIDIA)
https://arxiv.org/abs/1812.04948
The dataset consists of 70,000 high-quality PNG images at 1024×1024 resolution and contains considerable variation in terms of age, ethnicity and image background. It also has good coverage of accessories such as eyeglasses, sunglasses, hats, etc. The images were crawled from Flickr, thus inheriting all the biases of that website, and automatically aligned and cropped using dlib. Only images under permissive licenses were collected. Various automatic filters were used to prune the set, and finally Amazon Mechanical Turk was used to remove the occasional statues, paintings, or photos of photos.
Please note that this dataset is not intended for, and should not be used for, development or improvement of facial recognition technologies. For business inquiries, please visit our website and submit the form: NVIDIA Research Licensing
Licenses
The individual images were published in Flickr by their respective authors under either Creative Commons BY 2.0, Creative Commons BY-NC 2.0, Public Domain Mark 1.0, Public Domain CC0 1.0, or U.S. Government Works license. All of these licenses allow free use, redistribution, and adaptation for non-commercial purposes. However, some of them require giving appropriate credit to the original author, as well as indicating any changes that were made to the images. The license and original author of each image are indicated in the metadata.
- https://creativecommons.org/licenses/by/2.0/
- https://creativecommons.org/licenses/by-nc/2.0/
- https://creativecommons.org/publicdomain/mark/1.0/
- https://creativecommons.org/publicdomain/zero/1.0/
- http://www.usa.gov/copyright.shtml
The dataset itself (including JSON metadata, download script, and documentation) is made available under Creative Commons BY-NC-SA 4.0 license by NVIDIA Corporation. You can use, redistribute, and adapt it for non-commercial purposes, as long as you (a) give appropriate credit by citing our paper, (b) indicate any changes that you've made, and (c) distribute any derivative works under the same license.
- Downloads last month
- 253