Datasets:

ArXiv:
License:
orionweller commited on
Commit
13b1b43
·
verified ·
1 Parent(s): 1bb000e

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  2. train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  3. train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  4. train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  5. train/dclm-filtered_sampled-ext/split_10563-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  6. train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  7. train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  8. train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  9. train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  10. train/dclm-filtered_sampled-ext/split_11806-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  11. train/dclm-filtered_sampled-ext/split_12769-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  12. train/dclm-filtered_sampled-ext/split_14188-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  13. train/dclm-filtered_sampled-ext/split_14805-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  14. train/dclm-filtered_sampled-ext/split_14805-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  15. train/dclm-filtered_sampled-ext/split_14805-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  16. train/dclm-filtered_sampled-ext/split_1499-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  17. train/dclm-filtered_sampled-ext/split_15095-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  18. train/dclm-filtered_sampled-ext/split_15095-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  19. train/dclm-filtered_sampled-ext/split_15095-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  20. train/dclm-filtered_sampled-ext/split_15656-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  21. train/dclm-filtered_sampled-ext/split_15656-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  22. train/dclm-filtered_sampled-ext/split_15656-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  23. train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  24. train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  25. train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  26. train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  27. train/dclm-filtered_sampled-ext/split_16819-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  28. train/dclm-filtered_sampled-ext/split_16819-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  29. train/dclm-filtered_sampled-ext/split_16819-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  30. train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  31. train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  32. train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  33. train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  34. train/dclm-filtered_sampled-ext/split_17255-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/dclm-filtered_sampled-ext/split_17255-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  36. train/dclm-filtered_sampled-ext/split_17255-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  37. train/dclm-filtered_sampled-ext/split_18246-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  38. train/dclm-filtered_sampled-ext/split_18690-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  39. train/dclm-filtered_sampled-ext/split_18690-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  40. train/dclm-filtered_sampled-ext/split_18690-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  41. train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  42. train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  43. train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  44. train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  45. train/dclm-filtered_sampled-ext/split_19208-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  46. train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  47. train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  48. train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  49. train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  50. train/dclm-filtered_sampled-ext/split_19944-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106997, "hashes": {}}, "samples": 12626, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27862609, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103025, "hashes": {}}, "samples": 12638, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27752315, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 30577695, "hashes": {}}, "samples": 5525, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 12622768, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:918f1088170803cfb14ab02f8e45877b226d9dfaccf146ef545fa24cb823c379
3
+ size 30577695
train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 40532110, "total_tokens_skipped": 68, "percentiles": {"0th": 37, "10th": 227, "20th": 347, "30th": 458, "40th": 591, "50th": 743, "60th": 938, "70th": 1223, "80th": 1712, "90th": 3014, "95th": 5138, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_10185-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_10563-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed027f7e36e521f513242192019525e95170926e1232dcc66aa2e76a1b6ffc6d
3
+ size 16907480
train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103746, "hashes": {}}, "samples": 13071, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27814740, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107278, "hashes": {}}, "samples": 13932, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27956578, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 17307878, "hashes": {}}, "samples": 3579, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7222014, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:532075c7a3ce4d15c0d4898860f8745b3cae7c6e47ab56908aeddfe5c50d37f6
3
+ size 17307878
train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37201230, "total_tokens_skipped": 102, "percentiles": {"0th": 41, "10th": 231, "20th": 359, "30th": 483, "40th": 610, "50th": 746, "60th": 924, "70th": 1169, "80th": 1564, "90th": 2563, "95th": 4122, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_11260-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_11806-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e47567d543c5fa59ea775097e85e5e9dbd1b6ee56bfb8cf70ae165bb2c380bd
3
+ size 25188590
train/dclm-filtered_sampled-ext/split_12769-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a57dc39271a853010c743b53fc645be20513b1bbcc2537ec03fcc848acb2914
3
+ size 20765921
train/dclm-filtered_sampled-ext/split_14188-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b658fd2b6bf2a4c518865cf06a77ccc6675bc67c3478db7d30a9a800fab9150c
3
+ size 25324028
train/dclm-filtered_sampled-ext/split_14805-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67098979, "hashes": {}}, "samples": 13670, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 28004124, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106588, "hashes": {}}, "samples": 13689, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27894645, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 16822938, "hashes": {}}, "samples": 3336, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6970883, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_14805-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37100959, "total_tokens_skipped": 67, "percentiles": {"0th": 43, "10th": 222, "20th": 330, "30th": 438, "40th": 560, "50th": 699, "60th": 878, "70th": 1125, "80th": 1527, "90th": 2618, "95th": 4428, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_14805-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_1499-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7de6ed53ee74054c75bb973b1aa4aa3cff1ed0122dbcf63161750e77e0f01e32
3
+ size 16124502
train/dclm-filtered_sampled-ext/split_15095-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67081435, "hashes": {}}, "samples": 12760, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27893765, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67092202, "hashes": {}}, "samples": 12927, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27703100, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 26967519, "hashes": {}}, "samples": 5064, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11080679, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_15095-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 39619318, "total_tokens_skipped": 141, "percentiles": {"0th": 35, "10th": 220, "20th": 333, "30th": 451, "40th": 584, "50th": 736, "60th": 929, "70th": 1195, "80th": 1675, "90th": 2902, "95th": 4952, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_15095-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_15656-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67081560, "hashes": {}}, "samples": 12886, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27875608, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67087034, "hashes": {}}, "samples": 13596, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27972212, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 21232727, "hashes": {}}, "samples": 4200, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8870914, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_15656-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38199750, "total_tokens_skipped": 37, "percentiles": {"0th": 46, "10th": 223, "20th": 337, "30th": 448, "40th": 577, "50th": 720, "60th": 904, "70th": 1160, "80th": 1605, "90th": 2736, "95th": 4682, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_15656-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108214, "hashes": {}}, "samples": 13987, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27909295, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107264, "hashes": {}}, "samples": 13590, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27893300, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 15033961, "hashes": {}}, "samples": 2858, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6244638, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad2a1a7aa5dffd5c5251cecefdd21441fada21d73efcde300b27bef1fb37487c
3
+ size 15033961
train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36633987, "total_tokens_skipped": 122, "percentiles": {"0th": 30, "10th": 238, "20th": 367, "30th": 501, "40th": 635, "50th": 780, "60th": 965, "70th": 1212, "80th": 1594, "90th": 2478, "95th": 3871, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_1599-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_16819-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103753, "hashes": {}}, "samples": 13934, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 28010148, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67094892, "hashes": {}}, "samples": 13669, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28002149, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 15271770, "hashes": {}}, "samples": 2974, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6374656, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_16819-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36698503, "total_tokens_skipped": 57, "percentiles": {"0th": 30, "10th": 230, "20th": 354, "30th": 477, "40th": 603, "50th": 738, "60th": 910, "70th": 1152, "80th": 1539, "90th": 2499, "95th": 4056, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_16819-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108193, "hashes": {}}, "samples": 12531, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27723593, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67104302, "hashes": {}}, "samples": 13561, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27995028, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 22753141, "hashes": {}}, "samples": 4602, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9507395, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b482359943aa54fe3d2ce6b448f5f6b2ec2d760bb6456b205e50d774d33de59
3
+ size 22753141
train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38582004, "total_tokens_skipped": 89, "percentiles": {"0th": 48, "10th": 222, "20th": 338, "30th": 455, "40th": 582, "50th": 725, "60th": 908, "70th": 1168, "80th": 1630, "90th": 2788, "95th": 4662, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_17114-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_17255-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104260, "hashes": {}}, "samples": 12345, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27677436, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102232, "hashes": {}}, "samples": 13813, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27959779, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 21345700, "hashes": {}}, "samples": 4461, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8910107, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_17255-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38219888, "total_tokens_skipped": 116, "percentiles": {"0th": 36, "10th": 227, "20th": 350, "30th": 472, "40th": 600, "50th": 752, "60th": 934, "70th": 1192, "80th": 1619, "90th": 2697, "95th": 4439, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_17255-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_18246-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abea4179ad59fb66c24c7f7d9f6891a7500538d2c7f983757844cba739f44f4e
3
+ size 14089854
train/dclm-filtered_sampled-ext/split_18690-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105937, "hashes": {}}, "samples": 12672, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27824829, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105968, "hashes": {}}, "samples": 13029, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27889814, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 26140771, "hashes": {}}, "samples": 4995, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 10813236, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_18690-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 39427963, "total_tokens_skipped": 52, "percentiles": {"0th": 32, "10th": 225, "20th": 340, "30th": 459, "40th": 588, "50th": 733, "60th": 924, "70th": 1192, "80th": 1670, "90th": 2926, "95th": 4802, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_18690-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108620, "hashes": {}}, "samples": 13803, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27947037, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108657, "hashes": {}}, "samples": 13616, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28083621, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 16991511, "hashes": {}}, "samples": 3164, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7098380, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a84b3178e505d3280475b0d50be26d075661af548aa891dfd2c20a190c53bb2
3
+ size 16991511
train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37145702, "total_tokens_skipped": 60, "percentiles": {"0th": 42, "10th": 227, "20th": 347, "30th": 461, "40th": 587, "50th": 730, "60th": 908, "70th": 1157, "80th": 1572, "90th": 2593, "95th": 4276, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_19040-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_19208-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:508c7ba4c134682420328a57538a33e6b2b656232ee7f2ae6b9e110f9996a87d
3
+ size 23492380
train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105647, "hashes": {}}, "samples": 13370, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27910733, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105558, "hashes": {}}, "samples": 14452, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28010816, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 12211189, "hashes": {}}, "samples": 2606, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5079503, "hashes": {}}}], "version": 2}
train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d374923ffa70e226a6bb3ba8414131b3a8b0f5784da45a209faebc14a89d917
3
+ size 12211189
train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 35924334, "total_tokens_skipped": 62, "percentiles": {"0th": 39, "10th": 236, "20th": 372, "30th": 496, "40th": 623, "50th": 764, "60th": 938, "70th": 1173, "80th": 1542, "90th": 2412, "95th": 3744, "99th": 8190, "100th": 8191}}
train/dclm-filtered_sampled-ext/split_1949-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered_sampled-ext/split_19944-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67085888, "hashes": {}}, "samples": 12813, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27803039, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103417, "hashes": {}}, "samples": 13884, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28069702, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 20469831, "hashes": {}}, "samples": 3998, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8464520, "hashes": {}}}], "version": 2}