Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00043-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00043-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 75516347, "total_tokens_skipped": 45, "percentiles": {"0th": 65, "10th": 112, "20th": 149, "30th": 215, "40th": 305, "50th": 388, "60th": 491, "70th": 680, "80th": 1022, "90th": 1879, "95th": 3329, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67081525, "hashes": {}}, "samples": 22790, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6207987, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67100039, "hashes": {}}, "samples": 18076, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6990816, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67099939, "hashes": {}}, "samples": 17413, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8297473, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108693, "hashes": {}}, "samples": 18459, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6590103, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 32382722, "hashes": {}}, "samples": 10263, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 2974991, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 73808194, "total_tokens_skipped": 240, "percentiles": {"0th": 66, "10th": 113, "20th": 152, "30th": 226, "40th": 314, "50th": 396, "60th": 499, "70th": 676, "80th": 987, "90th": 1781, "95th": 3372, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67091358, "hashes": {}}, "samples": 20575, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8170298, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108626, "hashes": {}}, "samples": 16983, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9581367, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67086841, "hashes": {}}, "samples": 19049, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8423030, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 49670599, "hashes": {}}, "samples": 13742, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6379809, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 61619254, "total_tokens_skipped": 0, "percentiles": {"0th": 66, "10th": 113, "20th": 155, "30th": 234, "40th": 323, "50th": 405, "60th": 516, "70th": 705, "80th": 1045, "90th": 1893, "95th": 3513, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105540, "hashes": {}}, "samples": 15236, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15170287, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67086032, "hashes": {}}, "samples": 15839, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14879126, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 27635554, "hashes": {}}, "samples": 6093, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6513686, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 39864695, "total_tokens_skipped": 38, "percentiles": {"0th": 66, "10th": 116, "20th": 168, "30th": 259, "40th": 350, "50th": 448, "60th": 611, "70th": 871, "80th": 1327, "90th": 2482, "95th": 5152, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105507, "hashes": {}}, "samples": 13739, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 21046836, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 53963598, "hashes": {}}, "samples": 11675, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 17189004, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 29862374, "total_tokens_skipped": 14, "percentiles": {"0th": 69, "10th": 116, "20th": 168, "30th": 267, "40th": 373, "50th": 498, "60th": 698, "70th": 986, "80th": 1479, "90th": 2822, "95th": 5825, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107296, "hashes": {}}, "samples": 17752, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9764320, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108471, "hashes": {}}, "samples": 17442, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8884859, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67080349, "hashes": {}}, "samples": 19575, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8830837, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 47886519, "hashes": {}}, "samples": 10602, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6848559, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 61254653, "total_tokens_skipped": 157, "percentiles": {"0th": 64, "10th": 115, "20th": 158, "30th": 241, "40th": 332, "50th": 415, "60th": 542, "70th": 752, "80th": 1129, "90th": 2034, "95th": 3991, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67100017, "hashes": {}}, "samples": 18057, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8716297, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67101111, "hashes": {}}, "samples": 18925, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9137817, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67100387, "hashes": {}}, "samples": 18203, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9701344, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 37520892, "hashes": {}}, "samples": 9745, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5506663, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 58671735, "total_tokens_skipped": 101, "percentiles": {"0th": 68, "10th": 113, "20th": 156, "30th": 233, "40th": 325, "50th": 410, "60th": 525, "70th": 726, "80th": 1087, "90th": 2005, "95th": 3632, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67091312, "hashes": {}}, "samples": 17415, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12580423, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106581, "hashes": {}}, "samples": 16348, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12999129, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 17661492, "hashes": {}}, "samples": 5180, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 2899216, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 37344672, "total_tokens_skipped": 40, "percentiles": {"0th": 68, "10th": 115, "20th": 158, "30th": 241, "40th": 333, "50th": 425, "60th": 551, "70th": 791, "80th": 1189, "90th": 2152, "95th": 3895, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103575, "hashes": {}}, "samples": 14865, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16471315, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 37018390, "hashes": {}}, "samples": 8348, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9162530, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 25660681, "total_tokens_skipped": 2, "percentiles": {"0th": 67, "10th": 117, "20th": 168, "30th": 261, "40th": 355, "50th": 460, "60th": 642, "70th": 909, "80th": 1356, "90th": 2570, "95th": 5415, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67099471, "hashes": {}}, "samples": 13913, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 22723122, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 46162182, "hashes": {}}, "samples": 9122, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15920822, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 27948380, "total_tokens_skipped": 60, "percentiles": {"0th": 68, "10th": 118, "20th": 173, "30th": 274, "40th": 381, "50th": 516, "60th": 736, "70th": 1027, "80th": 1535, "90th": 2971, "95th": 6363, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107074, "hashes": {}}, "samples": 13607, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 23306956, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 44069078, "hashes": {}}, "samples": 8912, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15561485, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 27435221, "total_tokens_skipped": 104, "percentiles": {"0th": 69, "10th": 118, "20th": 173, "30th": 273, "40th": 380, "50th": 517, "60th": 732, "70th": 1027, "80th": 1525, "90th": 2938, "95th": 6362, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67098346, "hashes": {}}, "samples": 19347, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8431379, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106830, "hashes": {}}, "samples": 20092, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7772882, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107874, "hashes": {}}, "samples": 21220, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7358710, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 62569233, "hashes": {}}, "samples": 15032, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 8139372, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 64765602, "total_tokens_skipped": 80, "percentiles": {"0th": 66, "10th": 112, "20th": 152, "30th": 223, "40th": 313, "50th": 398, "60th": 513, "70th": 713, "80th": 1042, "90th": 1838, "95th": 3278, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107996, "hashes": {}}, "samples": 17381, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8716511, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107325, "hashes": {}}, "samples": 22163, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6850368, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67104597, "hashes": {}}, "samples": 17303, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8996376, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67099438, "hashes": {}}, "samples": 20324, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6913533, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 1742807, "hashes": {}}, "samples": 464, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 204270, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10ce9f3cdb649916e1d985f0545613d1cd4ed5df5ff614af756436cee9ba04b1
|
3 |
+
size 1742807
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 66304512, "total_tokens_skipped": 0, "percentiles": {"0th": 66, "10th": 112, "20th": 151, "30th": 225, "40th": 317, "50th": 400, "60th": 505, "70th": 693, "80th": 1014, "90th": 1816, "95th": 3331, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102032, "hashes": {}}, "samples": 17466, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9458569, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67096744, "hashes": {}}, "samples": 18052, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10194958, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108576, "hashes": {}}, "samples": 18346, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 10075339, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 27170511, "hashes": {}}, "samples": 6117, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 4681631, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 56164332, "total_tokens_skipped": 56, "percentiles": {"0th": 65, "10th": 115, "20th": 158, "30th": 238, "40th": 326, "50th": 412, "60th": 535, "70th": 755, "80th": 1119, "90th": 2087, "95th": 3923, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67095795, "hashes": {}}, "samples": 17336, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9958582, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103407, "hashes": {}}, "samples": 16959, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9911540, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67097608, "hashes": {}}, "samples": 17815, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9344200, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 39908929, "hashes": {}}, "samples": 10569, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5285739, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 59303280, "total_tokens_skipped": 25, "percentiles": {"0th": 68, "10th": 115, "20th": 160, "30th": 243, "40th": 331, "50th": 418, "60th": 536, "70th": 749, "80th": 1101, "90th": 2021, "95th": 4151, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67092275, "hashes": {}}, "samples": 16094, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13158294, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108091, "hashes": {}}, "samples": 16700, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13673454, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 10518914, "hashes": {}}, "samples": 2948, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 2024476, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0239556bb02a121bdd09a3e0b6140f71aa2356189c57ce844ca61feca785e940
|
3 |
+
size 10518914
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 35610489, "total_tokens_skipped": 60, "percentiles": {"0th": 66, "10th": 116, "20th": 165, "30th": 253, "40th": 342, "50th": 434, "60th": 577, "70th": 809, "80th": 1214, "90th": 2213, "95th": 4420, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00043-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104966, "hashes": {}}, "samples": 13460, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 24115162, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 39465326, "hashes": {}}, "samples": 8044, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14482601, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00043-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 26299942, "total_tokens_skipped": 36, "percentiles": {"0th": 69, "10th": 118, "20th": 172, "30th": 271, "40th": 381, "50th": 514, "60th": 726, "70th": 1027, "80th": 1526, "90th": 2987, "95th": 6366, "99th": 8190, "100th": 8191}}
|