Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0054-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0054-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0054-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0154-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0154-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0154-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0063-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0063-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0063-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0161-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0161-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0161-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0044-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0044-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0044-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106501, "hashes": {}}, "samples": 24787, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6101443, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107160, "hashes": {}}, "samples": 25094, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6040345, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108089, "hashes": {}}, "samples": 24374, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6070691, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 56808504, "hashes": {}}, "samples": 19250, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5077176, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 63043463, "total_tokens_skipped": 0, "percentiles": {"0th": 67, "10th": 130, "20th": 188, "30th": 249, "40th": 326, "50th": 418, "60th": 534, "70th": 681, "80th": 910, "90th": 1378, "95th": 2008, "99th": 4806, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0054-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108121, "hashes": {}}, "samples": 22905, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6576653, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105409, "hashes": {}}, "samples": 23557, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6596608, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108501, "hashes": {}}, "samples": 23014, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6590416, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 42035227, "hashes": {}}, "samples": 14611, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 4416541, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0054-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 59499562, "total_tokens_skipped": 0, "percentiles": {"0th": 68, "10th": 138, "20th": 204, "30th": 270, "40th": 355, "50th": 454, "60th": 580, "70th": 746, "80th": 980, "90th": 1451, "95th": 2071, "99th": 4535, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0054-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107636, "hashes": {}}, "samples": 22740, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7441294, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67099025, "hashes": {}}, "samples": 14303, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6550878, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67104057, "hashes": {}}, "samples": 22763, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6160643, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 64503002, "hashes": {}}, "samples": 23974, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5984513, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 65118300, "total_tokens_skipped": 95, "percentiles": {"0th": 67, "10th": 143, "20th": 208, "30th": 280, "40th": 363, "50th": 468, "60th": 595, "70th": 774, "80th": 1032, "90th": 1581, "95th": 2415, "99th": 6194, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67086887, "hashes": {}}, "samples": 23497, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6863981, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107687, "hashes": {}}, "samples": 22765, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7145456, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107635, "hashes": {}}, "samples": 22076, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7563982, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 23893626, "hashes": {}}, "samples": 8043, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 2805203, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 55081763, "total_tokens_skipped": 0, "percentiles": {"0th": 65, "10th": 141, "20th": 207, "30th": 280, "40th": 366, "50th": 461, "60th": 592, "70th": 757, "80th": 996, "90th": 1489, "95th": 2121, "99th": 4577, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0154-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107974, "hashes": {}}, "samples": 21924, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6596600, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105159, "hashes": {}}, "samples": 25300, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6418257, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107298, "hashes": {}}, "samples": 25451, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6389347, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 40162597, "hashes": {}}, "samples": 14611, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 3906204, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0154-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 58980351, "total_tokens_skipped": 0, "percentiles": {"0th": 68, "10th": 136, "20th": 195, "30th": 260, "40th": 332, "50th": 423, "60th": 536, "70th": 688, "80th": 911, "90th": 1369, "95th": 1922, "99th": 4849, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0154-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108456, "hashes": {}}, "samples": 24834, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6587535, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107355, "hashes": {}}, "samples": 26156, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6505882, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67105297, "hashes": {}}, "samples": 24057, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6550481, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 27639434, "hashes": {}}, "samples": 10140, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 2851144, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 55883248, "total_tokens_skipped": 120, "percentiles": {"0th": 66, "10th": 137, "20th": 194, "30th": 258, "40th": 329, "50th": 418, "60th": 527, "70th": 675, "80th": 897, "90th": 1316, "95th": 1819, "99th": 4596, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107926, "hashes": {}}, "samples": 20353, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7801098, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107856, "hashes": {}}, "samples": 26664, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6719841, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107600, "hashes": {}}, "samples": 25487, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6910913, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 13521101, "hashes": {}}, "samples": 5084, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 1498518, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d923d43e4fc7c8c504560609328bb2ff9261b92bd66831afe03ab377cd504aa3
|
3 |
+
size 13521101
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 52475060, "total_tokens_skipped": 0, "percentiles": {"0th": 62, "10th": 139, "20th": 198, "30th": 264, "40th": 338, "50th": 429, "60th": 540, "70th": 691, "80th": 921, "90th": 1384, "95th": 1954, "99th": 4493, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0065-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106389, "hashes": {}}, "samples": 23155, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7988127, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105790, "hashes": {}}, "samples": 19794, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8204407, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67105239, "hashes": {}}, "samples": 25279, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6782714, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 12142901, "hashes": {}}, "samples": 4568, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 1217617, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0b4a028741da7fc1350888bfe4ad46d45ef1927a8e00f72255dd5a538538d0b7
|
3 |
+
size 12142901
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 52205110, "total_tokens_skipped": 0, "percentiles": {"0th": 64, "10th": 141, "20th": 204, "30th": 275, "40th": 357, "50th": 455, "60th": 575, "70th": 735, "80th": 985, "90th": 1452, "95th": 2087, "99th": 5083, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0068-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67093505, "hashes": {}}, "samples": 23922, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7845265, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67097335, "hashes": {}}, "samples": 22870, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8733348, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108243, "hashes": {}}, "samples": 21290, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7091700, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 5178060, "hashes": {}}, "samples": 1930, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 545548, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f903711cbf764216ae0c6e04bb9ad6255737667bbcba3ba427f918d1f86159d7
|
3 |
+
size 5178060
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 50503554, "total_tokens_skipped": 160, "percentiles": {"0th": 70, "10th": 144, "20th": 208, "30th": 281, "40th": 362, "50th": 458, "60th": 580, "70th": 747, "80th": 991, "90th": 1469, "95th": 2067, "99th": 4794, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103074, "hashes": {}}, "samples": 23625, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9793703, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108846, "hashes": {}}, "samples": 23470, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9374676, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 25446082, "hashes": {}}, "samples": 9851, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3516837, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:98bbf34ab7a81661188171f964489dc223cfa31fab39fdd43d39b2b260ac9222
|
3 |
+
size 25446082
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 39007182, "total_tokens_skipped": 0, "percentiles": {"0th": 68, "10th": 141, "20th": 204, "30th": 270, "40th": 346, "50th": 437, "60th": 551, "70th": 705, "80th": 938, "90th": 1385, "95th": 1975, "99th": 4580, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0063-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104687, "hashes": {}}, "samples": 21359, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10144280, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106922, "hashes": {}}, "samples": 24272, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9116898, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 26384709, "hashes": {}}, "samples": 9211, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 4173056, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0063-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 39275048, "total_tokens_skipped": 18, "percentiles": {"0th": 68, "10th": 144, "20th": 211, "30th": 281, "40th": 359, "50th": 453, "60th": 569, "70th": 727, "80th": 962, "90th": 1417, "95th": 2026, "99th": 5303, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0063-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108634, "hashes": {}}, "samples": 24186, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10953055, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108318, "hashes": {}}, "samples": 22298, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11604866, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 3554332, "hashes": {}}, "samples": 1296, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 495897, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:43a6873b7b21612eae3ba74e7bc134fe37b2a3a2979772cec759a465175037d0
|
3 |
+
size 3554332
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 33681289, "total_tokens_skipped": 0, "percentiles": {"0th": 67, "10th": 145, "20th": 213, "30th": 282, "40th": 363, "50th": 461, "60th": 572, "70th": 724, "80th": 959, "90th": 1418, "95th": 2037, "99th": 4611, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0118-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0161-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107175, "hashes": {}}, "samples": 24139, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11968365, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 58659755, "hashes": {}}, "samples": 20272, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10271260, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0161-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 30733983, "total_tokens_skipped": 0, "percentiles": {"0th": 68, "10th": 144, "20th": 210, "30th": 280, "40th": 356, "50th": 446, "60th": 560, "70th": 709, "80th": 935, "90th": 1379, "95th": 1943, "99th": 4685, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0161-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108303, "hashes": {}}, "samples": 23372, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14825641, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 34018411, "hashes": {}}, "samples": 11545, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7487692, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 24725216, "total_tokens_skipped": 63, "percentiles": {"0th": 67, "10th": 144, "20th": 210, "30th": 278, "40th": 350, "50th": 446, "60th": 559, "70th": 714, "80th": 939, "90th": 1409, "95th": 2031, "99th": 5297, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0044-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107606, "hashes": {}}, "samples": 24048, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16320726, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 23699770, "hashes": {}}, "samples": 8492, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5422592, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0044-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 22183268, "total_tokens_skipped": 0, "percentiles": {"0th": 66, "10th": 144, "20th": 210, "30th": 279, "40th": 352, "50th": 446, "60th": 554, "70th": 705, "80th": 926, "90th": 1374, "95th": 1953, "99th": 4508, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0044-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|