Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0024-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0031-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0031-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0031-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0078-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0078-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0078-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0093-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0093-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0093-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0121-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0121-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0121-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0207-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0207-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0207-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0211-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0211-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0211-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0213-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0213-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0213-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0109-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0109-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0109-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0174-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0174-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0174-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0181-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0024-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107588, "hashes": {}}, "samples": 18535, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5549084, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105770, "hashes": {}}, "samples": 17882, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5991289, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67104895, "hashes": {}}, "samples": 17577, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6509970, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108674, "hashes": {}}, "samples": 16836, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6898575, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 9720276, "hashes": {}}, "samples": 2344, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 939404, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0031-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107555, "hashes": {}}, "samples": 17917, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5232005, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67099069, "hashes": {}}, "samples": 18696, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5662903, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67106649, "hashes": {}}, "samples": 18081, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6278403, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107513, "hashes": {}}, "samples": 17564, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6619206, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 20678261, "hashes": {}}, "samples": 4891, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 2156308, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0031-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 71044244, "total_tokens_skipped": 50, "percentiles": {"0th": 86, "10th": 190, "20th": 247, "30th": 306, "40th": 373, "50th": 462, "60th": 593, "70th": 814, "80th": 1202, "90th": 2015, "95th": 3285, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0031-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0078-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108154, "hashes": {}}, "samples": 19863, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6344559, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108745, "hashes": {}}, "samples": 18847, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7234434, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67104463, "hashes": {}}, "samples": 17823, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6840150, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 44991656, "hashes": {}}, "samples": 13247, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 3937818, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0078-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 60465408, "total_tokens_skipped": 8, "percentiles": {"0th": 86, "10th": 190, "20th": 244, "30th": 301, "40th": 365, "50th": 450, "60th": 568, "70th": 768, "80th": 1112, "90th": 1869, "95th": 2998, "99th": 7691, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0078-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0093-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108515, "hashes": {}}, "samples": 19041, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7724216, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108362, "hashes": {}}, "samples": 18286, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6024220, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107681, "hashes": {}}, "samples": 19885, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6900204, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 30131193, "hashes": {}}, "samples": 8493, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 3237433, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0093-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 56815944, "total_tokens_skipped": 0, "percentiles": {"0th": 91, "10th": 192, "20th": 249, "30th": 307, "40th": 375, "50th": 460, "60th": 579, "70th": 761, "80th": 1107, "90th": 1888, "95th": 2892, "99th": 7636, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0093-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0121-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108513, "hashes": {}}, "samples": 19465, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6785417, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107635, "hashes": {}}, "samples": 19483, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7641437, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107725, "hashes": {}}, "samples": 19366, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7420215, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 20672403, "hashes": {}}, "samples": 6670, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 2012835, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0121-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 54462833, "total_tokens_skipped": 0, "percentiles": {"0th": 84, "10th": 187, "20th": 241, "30th": 297, "40th": 358, "50th": 436, "60th": 545, "70th": 727, "80th": 1061, "90th": 1787, "95th": 2830, "99th": 8094, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0121-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107893, "hashes": {}}, "samples": 22099, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8718639, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107625, "hashes": {}}, "samples": 21121, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8749094, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 32567629, "hashes": {}}, "samples": 8895, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 4831271, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 40864648, "total_tokens_skipped": 0, "percentiles": {"0th": 88, "10th": 190, "20th": 245, "30th": 299, "40th": 363, "50th": 436, "60th": 538, "70th": 692, "80th": 966, "90th": 1640, "95th": 2565, "99th": 6405, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0207-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107306, "hashes": {}}, "samples": 21328, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9483611, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67097754, "hashes": {}}, "samples": 20593, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9645543, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 22243813, "hashes": {}}, "samples": 7402, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 2597975, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0207-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 38325613, "total_tokens_skipped": 0, "percentiles": {"0th": 84, "10th": 189, "20th": 244, "30th": 296, "40th": 357, "50th": 430, "60th": 532, "70th": 680, "80th": 972, "90th": 1625, "95th": 2521, "99th": 6203, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0207-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0211-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105911, "hashes": {}}, "samples": 21320, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9395231, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103212, "hashes": {}}, "samples": 21044, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9234039, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 23202880, "hashes": {}}, "samples": 6419, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3373433, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0211-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 38575123, "total_tokens_skipped": 0, "percentiles": {"0th": 92, "10th": 189, "20th": 240, "30th": 295, "40th": 355, "50th": 433, "60th": 530, "70th": 682, "80th": 975, "90th": 1649, "95th": 2644, "99th": 6755, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0211-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0213-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107316, "hashes": {}}, "samples": 19478, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9538356, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106525, "hashes": {}}, "samples": 21389, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9821196, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 19642943, "hashes": {}}, "samples": 6883, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 2806453, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0213-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 37702714, "total_tokens_skipped": 6, "percentiles": {"0th": 86, "10th": 190, "20th": 244, "30th": 295, "40th": 354, "50th": 425, "60th": 523, "70th": 668, "80th": 948, "90th": 1628, "95th": 2596, "99th": 7720, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0213-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104274, "hashes": {}}, "samples": 21131, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9645501, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106307, "hashes": {}}, "samples": 22219, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9645955, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 13024614, "hashes": {}}, "samples": 4485, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 1959372, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14f72be345dde1124774ecc565dccf97b6dda6abe48609d5bce9aa0eec8107a7
|
3 |
+
size 13024614
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 36045971, "total_tokens_skipped": 0, "percentiles": {"0th": 86, "10th": 191, "20th": 243, "30th": 295, "40th": 352, "50th": 423, "60th": 522, "70th": 672, "80th": 935, "90th": 1561, "95th": 2484, "99th": 5820, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104459, "hashes": {}}, "samples": 21055, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10241871, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67104444, "hashes": {}}, "samples": 20703, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10015006, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 14935336, "hashes": {}}, "samples": 4973, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 2118764, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a58151fbf5c4025a1ebccf61bcbda37200f94167510c02524556f1be90d1a289
|
3 |
+
size 14935336
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 36540838, "total_tokens_skipped": 0, "percentiles": {"0th": 89, "10th": 189, "20th": 243, "30th": 296, "40th": 356, "50th": 433, "60th": 536, "70th": 696, "80th": 983, "90th": 1621, "95th": 2575, "99th": 6398, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0009-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67083464, "hashes": {}}, "samples": 22406, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10593676, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67104676, "hashes": {}}, "samples": 20662, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10042804, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 3110134, "hashes": {}}, "samples": 910, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 527738, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bf1fd557498183db7a8785e1ebe5f70e7aca8461eb995559a73dbee4c715bdad
|
3 |
+
size 3110134
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 33623148, "total_tokens_skipped": 23, "percentiles": {"0th": 96, "10th": 191, "20th": 246, "30th": 299, "40th": 359, "50th": 431, "60th": 534, "70th": 680, "80th": 942, "90th": 1551, "95th": 2449, "99th": 6283, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0015-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107050, "hashes": {}}, "samples": 21033, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13639051, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 37736657, "hashes": {}}, "samples": 12374, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7759090, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 25678160, "total_tokens_skipped": 0, "percentiles": {"0th": 92, "10th": 191, "20th": 244, "30th": 299, "40th": 361, "50th": 432, "60th": 531, "70th": 670, "80th": 936, "90th": 1597, "95th": 2539, "99th": 6078, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105142, "hashes": {}}, "samples": 21854, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14705390, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 26558793, "hashes": {}}, "samples": 9041, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5851502, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 22923274, "total_tokens_skipped": 18, "percentiles": {"0th": 89, "10th": 192, "20th": 245, "30th": 299, "40th": 359, "50th": 430, "60th": 529, "70th": 668, "80th": 917, "90th": 1490, "95th": 2389, "99th": 5606, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0109-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67085153, "hashes": {}}, "samples": 22250, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14855280, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 25345635, "hashes": {}}, "samples": 8416, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5603893, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0109-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 22618625, "total_tokens_skipped": 20, "percentiles": {"0th": 90, "10th": 194, "20th": 246, "30th": 297, "40th": 357, "50th": 427, "60th": 525, "70th": 660, "80th": 904, "90th": 1506, "95th": 2358, "99th": 5516, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0109-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0174-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 65694952, "hashes": {}}, "samples": 21364, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 20419083, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0174-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 16083059, "total_tokens_skipped": 11, "percentiles": {"0th": 85, "10th": 190, "20th": 243, "30th": 295, "40th": 356, "50th": 428, "60th": 523, "70th": 661, "80th": 898, "90th": 1505, "95th": 2445, "99th": 6073, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0174-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0181-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 63595429, "hashes": {}}, "samples": 20314, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 20577136, "hashes": {}}}], "version": 2}
|