Datasets:

ArXiv:
License:
orionweller commited on
Commit
83cc406
·
verified ·
1 Parent(s): 7a2d05c

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  2. train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
  3. train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00006.mds +3 -0
  4. train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00011.mds +3 -0
  5. train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00013.mds +3 -0
  6. train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  7. train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  8. train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  9. train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  10. train/dclm-filtered-sampled-decay-10B/split_10841-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  11. train/dclm-filtered-sampled-decay-10B/split_10841-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  12. train/dclm-filtered-sampled-decay-10B/split_10841-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  13. train/dclm-filtered-sampled-decay-10B/split_10851-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  14. train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  15. train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  16. train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  17. train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  18. train/dclm-filtered-sampled-decay-10B/split_11056-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  19. train/dclm-filtered-sampled-decay-10B/split_11056-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  20. train/dclm-filtered-sampled-decay-10B/split_11056-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  21. train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  22. train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  23. train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  24. train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  25. train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  26. train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  27. train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  28. train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  29. train/dclm-filtered-sampled-decay-10B/split_11515-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  30. train/dclm-filtered-sampled-decay-10B/split_11515-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  31. train/dclm-filtered-sampled-decay-10B/split_11780-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  32. train/dclm-filtered-sampled-decay-10B/split_11780-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  33. train/dclm-filtered-sampled-decay-10B/split_11945-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  34. train/dclm-filtered-sampled-decay-10B/split_11945-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  35. train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  36. train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  37. train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  38. train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  39. train/dclm-filtered-sampled-decay-10B/split_12687-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  40. train/dclm-filtered-sampled-decay-10B/split_12687-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  41. train/dclm-filtered-sampled-decay-10B/split_12689-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  42. train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  43. train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  44. train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  45. train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  46. train/dclm-filtered-sampled-decay-10B/split_12720-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  47. train/dclm-filtered-sampled-decay-10B/split_12720-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  48. train/dclm-filtered-sampled-decay-10B/split_12755-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  49. train/dclm-filtered-sampled-decay-10B/split_12755-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  50. train/dclm-filtered-sampled-decay-10B/split_12755-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d15bf04ed484907e76a6e2100b9a81ad16af361244d1165d338cb180c05f4580
3
+ size 67101459
train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbdd5394e3b007eaecf9f60f846e7a1a594bc287ae444d35d68e79ba8dc349da
3
+ size 67102284
train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddd50678886ec09451562eda496fbc12addd9c5fe60a01948c664c1ef9c49fad
3
+ size 67099409
train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00011.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f8cbfd0a9ac95e1b3ab1a6f567d32d4347df847eaa7fb5f94729e7d4ef8d54c
3
+ size 67105522
train/arxiv-sampled-decay/arxiv_0075-tokenized-chunked-8192-512-32-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ad7909a8aaddb55cef6645c17b822d10b8a2deb93ce8e2bf6027af4e83f7c7e
3
+ size 67103317
train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67097256, "hashes": {}}, "samples": 13755, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27936468, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108656, "hashes": {}}, "samples": 13856, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28084805, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 14162697, "hashes": {}}, "samples": 2883, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5934166, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b5fd1dd2b55e09e432751d33d8e065f770d037005bf4dd8f4229018878c08f7
3
+ size 14162697
train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36420719, "total_tokens_skipped": 49, "percentiles": {"0th": 38, "10th": 240, "20th": 370, "30th": 491, "40th": 617, "50th": 755, "60th": 924, "70th": 1157, "80th": 1541, "90th": 2463, "95th": 3995, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_10033-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_10841-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67088242, "hashes": {}}, "samples": 12449, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27616839, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105272, "hashes": {}}, "samples": 13806, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28035953, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 21205748, "hashes": {}}, "samples": 4464, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8900032, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_10841-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38192422, "total_tokens_skipped": 42, "percentiles": {"0th": 43, "10th": 224, "20th": 339, "30th": 449, "40th": 573, "50th": 714, "60th": 896, "70th": 1143, "80th": 1583, "90th": 2723, "95th": 4693, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_10841-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_10851-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50fa592c257d47452a7c6d65781788525eb7c8ecfc9133aeb5bc4c2e066676a6
3
+ size 14080905
train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105187, "hashes": {}}, "samples": 13745, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27961834, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67098651, "hashes": {}}, "samples": 13684, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28041847, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 16245920, "hashes": {}}, "samples": 3214, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6780680, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c3eaf50b3d6e3866abf3d27594861fecc8e77660cca1907673c6c74ec335fcf
3
+ size 16245920
train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36954869, "total_tokens_skipped": 40, "percentiles": {"0th": 33, "10th": 224, "20th": 339, "30th": 448, "40th": 574, "50th": 709, "60th": 880, "70th": 1131, "80th": 1541, "90th": 2574, "95th": 4350, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_10940-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_11056-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103944, "hashes": {}}, "samples": 12726, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27674354, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108194, "hashes": {}}, "samples": 13512, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 28007571, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 21681501, "hashes": {}}, "samples": 4375, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9066645, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_11056-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38304227, "total_tokens_skipped": 52, "percentiles": {"0th": 43, "10th": 225, "20th": 353, "30th": 474, "40th": 607, "50th": 749, "60th": 934, "70th": 1192, "80th": 1625, "90th": 2708, "95th": 4441, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_11056-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107428, "hashes": {}}, "samples": 14458, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 28184463, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106876, "hashes": {}}, "samples": 13730, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27631737, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 11710165, "hashes": {}}, "samples": 2325, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 4790521, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b95f9f1c2b13fe5630308718de1e3a03db4213f789b5abb0e17995eda078d1d0
3
+ size 11710165
train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 35814752, "total_tokens_skipped": 30, "percentiles": {"0th": 35, "10th": 224, "20th": 342, "30th": 459, "40th": 581, "50th": 714, "60th": 884, "70th": 1123, "80th": 1521, "90th": 2483, "95th": 4014, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_11350-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108699, "hashes": {}}, "samples": 14210, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 28186817, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108343, "hashes": {}}, "samples": 13107, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27865530, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 17108638, "hashes": {}}, "samples": 3264, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7127285, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f02bb4cc5f8023ea9d7dc4ce642a51c57e2ed748da4f01464597d078b55360a4
3
+ size 17108638
train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37181197, "total_tokens_skipped": 24, "percentiles": {"0th": 38, "10th": 225, "20th": 338, "30th": 444, "40th": 570, "50th": 706, "60th": 892, "70th": 1144, "80th": 1594, "90th": 2682, "95th": 4365, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_11477-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_11515-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2b22b1a86004f85277f31028d440954bf36e2f2692f31f24140f9ff31ba14bd
3
+ size 67107145
train/dclm-filtered-sampled-decay-10B/split_11515-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8217b9eb166d3197537174906cb0a6b8a9b594135bda9bd355d3c47dcdcd3a6d
3
+ size 67108805
train/dclm-filtered-sampled-decay-10B/split_11780-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a92cea74c8743fb4a243c00da1b027223c251ebcd5515372646a5418bb2c7e5a
3
+ size 67098983
train/dclm-filtered-sampled-decay-10B/split_11780-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f97b62a04ed777c69c35f95989c32b4e74c5f2bd34ff583f3aa55a0d38ce9a1a
3
+ size 67108084
train/dclm-filtered-sampled-decay-10B/split_11945-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d48e2664d22370b6a8d31930ec9ff2fbde9baec06e06fc69a1aedbb49340753
3
+ size 67099724
train/dclm-filtered-sampled-decay-10B/split_11945-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24bad1d4d95e4a22dff3d11ea35ad8cce889414aa8746692c89058eea393ff64
3
+ size 67106823
train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103202, "hashes": {}}, "samples": 13268, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27832729, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67096892, "hashes": {}}, "samples": 13737, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27892788, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 17433994, "hashes": {}}, "samples": 3524, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7246728, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:733e092889a2358993e1d34b389ea584ffa80d2dab15178bc7edfccac46dfb0e
3
+ size 17433994
train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37231574, "total_tokens_skipped": 68, "percentiles": {"0th": 47, "10th": 238, "20th": 363, "30th": 491, "40th": 624, "50th": 772, "60th": 952, "70th": 1195, "80th": 1583, "90th": 2545, "95th": 4031, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_12648-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_12687-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88b9187edf732b31d27ad9bd3c344a0bb1798d92d1956aebeb7d1ff605c24cd7
3
+ size 67103720
train/dclm-filtered-sampled-decay-10B/split_12687-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eb51a6e54159e13aab5ac7d3993dbbca402ac4e25e37f4c13a4318d0da77a94
3
+ size 67105203
train/dclm-filtered-sampled-decay-10B/split_12689-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b51188d3a1fab60e4f04fabc9942c5736d8522c9a461f523ef98fed640a91d55
3
+ size 12135964
train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108216, "hashes": {}}, "samples": 13488, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 28047057, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67091571, "hashes": {}}, "samples": 13086, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27848776, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 21774679, "hashes": {}}, "samples": 4133, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9031746, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:439fc4256b6c872b3d0bd6d386dbc6797d0c8cdf2d47257f9cc1bc87860f4f9f
3
+ size 21774679
train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38341007, "total_tokens_skipped": 22, "percentiles": {"0th": 38, "10th": 231, "20th": 341, "30th": 457, "40th": 578, "50th": 721, "60th": 907, "70th": 1157, "80th": 1602, "90th": 2719, "95th": 4677, "99th": 8190, "100th": 8191}}
train/dclm-filtered-sampled-decay-10B/split_12702-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/dclm-filtered-sampled-decay-10B/split_12720-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad2f31a69ebfe7b677f187bd67e133d22856b64bd73809a3762be7b19b894706
3
+ size 67108578
train/dclm-filtered-sampled-decay-10B/split_12720-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c249f0fa3d25a6f045da34add1d6fda4486aad989e9c629c48fca5d195b23377
3
+ size 67096898
train/dclm-filtered-sampled-decay-10B/split_12755-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106105, "hashes": {}}, "samples": 14183, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27987192, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105413, "hashes": {}}, "samples": 13729, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 27926803, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 12489945, "hashes": {}}, "samples": 2561, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5200362, "hashes": {}}}], "version": 2}
train/dclm-filtered-sampled-decay-10B/split_12755-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5160012c932f0ea8894a117fc258fda745a14bd2b06820aee15db66a75603193
3
+ size 12489945
train/dclm-filtered-sampled-decay-10B/split_12755-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36006162, "total_tokens_skipped": 30, "percentiles": {"0th": 39, "10th": 235, "20th": 364, "30th": 492, "40th": 620, "50th": 759, "60th": 942, "70th": 1183, "80th": 1557, "90th": 2425, "95th": 3700, "99th": 8190, "100th": 8191}}