Datasets:

ArXiv:
License:
orionweller commited on
Commit
9fb191e
·
verified ·
1 Parent(s): 8b43eb6

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  2. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  3. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  4. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  5. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  6. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  7. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  8. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  9. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  10. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  11. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  12. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  13. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  14. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  15. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  16. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  17. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  18. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  19. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  20. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  21. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  22. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  23. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00101-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  24. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  25. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  26. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  27. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  28. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00132-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  29. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00132-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  30. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00132-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  31. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00147-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  32. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  33. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  34. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00186-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  35. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  36. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  37. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  38. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  39. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  40. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  41. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  42. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  43. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00250-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  44. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00266-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  45. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00266-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  46. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  47. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  48. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00341-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  49. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00341-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  50. train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00341-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c3b8a3c68e567360623853c0ea92335c970938e7d5db153e5f30355ef0f19d7
3
+ size 67108744
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69319b2b6fcec1712bcad1aa3c220cc697b35b1e7517dfa9a1b0d27cc8bd50ea
3
+ size 67107461
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f79d8bd016a128f366da39c37d6a2e0c7dbbcd8187a7bbc4869880b3be508c5
3
+ size 67107661
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1916c4fac3beca84b98edc60f4e1ab4c42958b53fb8c436374c83137f35e28b5
3
+ size 67104694
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0073a6b62c1bb77079a9df683af3a6b1c91e769f464d9449c7c5eb5946223f7b
3
+ size 67102121
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108599, "hashes": {}}, "samples": 15341, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7218946, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108579, "hashes": {}}, "samples": 18118, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7831358, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67106454, "hashes": {}}, "samples": 18144, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7740307, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107979, "hashes": {}}, "samples": 17503, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7926158, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 40205295, "hashes": {}}, "samples": 9337, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 4726216, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 75908582, "total_tokens_skipped": 0, "percentiles": {"0th": 61, "10th": 174, "20th": 232, "30th": 294, "40th": 367, "50th": 475, "60th": 664, "70th": 902, "80th": 1290, "90th": 2188, "95th": 3469, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101573, "hashes": {}}, "samples": 17114, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8331572, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105959, "hashes": {}}, "samples": 17870, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8579188, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67104812, "hashes": {}}, "samples": 18538, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8767967, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 61715221, "hashes": {}}, "samples": 15105, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 8258922, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 64663007, "total_tokens_skipped": 60, "percentiles": {"0th": 60, "10th": 169, "20th": 224, "30th": 286, "40th": 352, "50th": 443, "60th": 622, "70th": 854, "80th": 1251, "90th": 2173, "95th": 3489, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101799, "hashes": {}}, "samples": 17337, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8906809, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107825, "hashes": {}}, "samples": 17720, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9178944, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67103751, "hashes": {}}, "samples": 17895, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9466579, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 37893200, "hashes": {}}, "samples": 9814, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5707365, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 58801222, "total_tokens_skipped": 135, "percentiles": {"0th": 61, "10th": 166, "20th": 222, "30th": 282, "40th": 349, "50th": 440, "60th": 622, "70th": 867, "80th": 1274, "90th": 2145, "95th": 3360, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c12606a323de97187d43476da3151e7aa3dd35172a5c48c6169b66603df4d78
3
+ size 67108268
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d4901d7e0fbe33a950541c5051907d29c8c488978f6a2ba1c4e3bcc3113b697
3
+ size 67093974
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:718cf05d30d6cb75243de9ffb9e1ee209551dad08e4ad1bf68448e85e927c3a9
3
+ size 67107907
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00066-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92b10edcf754a3d808b5d0aa30e2916a1ebf0ba77f7d14f62636b7357cd90303
3
+ size 34659316
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aca5685dc6afcce506529346a0b5fdd3d1da0dbea89e2e64bc1230430abeb794
3
+ size 67107316
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02b7074f563c2471453c728c20fbcfbe1b83e60a7f57a1eb03f02875ff6a4f8b
3
+ size 67104191
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:207176a428d969bd36778f8856ef209cf48434301f35d680600d5c54190d1cec
3
+ size 67107631
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00071-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51dc6c338f617eb15c41a3bb7e8edbee0aeb90edf1946c1d1cb8be982ead12ca
3
+ size 67100217
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00101-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:339e08c76e4b38c185fc44169f3beaf455f2c17a24099fd2b6ef8af8b9658676
3
+ size 47155274
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108489, "hashes": {}}, "samples": 16359, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11409047, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105497, "hashes": {}}, "samples": 16650, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12934761, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108137, "hashes": {}}, "samples": 16352, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 12027144, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 16748496, "hashes": {}}, "samples": 4306, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 3322029, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aefc7fdac6849ff322e1afd55fe8dafd16353eab9758a8f66836dbef37e0fb31
3
+ size 16748496
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 53660902, "total_tokens_skipped": 258, "percentiles": {"0th": 60, "10th": 209, "20th": 285, "30th": 348, "40th": 420, "50th": 541, "60th": 708, "70th": 952, "80th": 1378, "90th": 2297, "95th": 3323, "99th": 7739, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00132-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67098184, "hashes": {}}, "samples": 17411, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12601318, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 33150061, "hashes": {}}, "samples": 7213, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6460152, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00132-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 24668905, "total_tokens_skipped": 0, "percentiles": {"0th": 61, "10th": 211, "20th": 292, "30th": 353, "40th": 423, "50th": 551, "60th": 718, "70th": 954, "80th": 1366, "90th": 2333, "95th": 3271, "99th": 7752, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00132-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00147-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b123a00bea1ca479206ea71e10d4d6710dad7b61337c3673a1af082f64645c8e
3
+ size 8781310
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5c0db2a3cca8964f2163fdbd04fe4348e9fffdfd2fce3b0c990a0e924b3b4aa
3
+ size 67104742
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d746595b36fa1080e3c60a45dc09112b664f45ad7f345ac35656c52eb333370f
3
+ size 45830545
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00186-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:354b50b51d86957d866d201624f2e3a2b70d19c0cefcf4b6f4460862221a6335
3
+ size 44491029
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67091751, "hashes": {}}, "samples": 18730, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13998342, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20179396, "hashes": {}}, "samples": 5925, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3806931, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b48cd45700d1fb5650bb7960c5cd680e8df4eae90b446c69c969092b5fa215ab
3
+ size 20179396
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21424219, "total_tokens_skipped": 0, "percentiles": {"0th": 61, "10th": 206, "20th": 279, "30th": 337, "40th": 400, "50th": 504, "60th": 656, "70th": 852, "80th": 1178, "90th": 1829, "95th": 2730, "99th": 5992, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00187-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108801, "hashes": {}}, "samples": 18957, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13412759, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107470, "hashes": {}}, "samples": 20144, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15390535, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 19471682, "hashes": {}}, "samples": 5820, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3802549, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9ab5c4aa5387b4718ead661dd38f5cda32e39eee8e2e0145c8c5ff839ba01bb
3
+ size 19471682
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37705327, "total_tokens_skipped": 25, "percentiles": {"0th": 57, "10th": 192, "20th": 256, "30th": 315, "40th": 375, "50th": 458, "60th": 601, "70th": 794, "80th": 1104, "90th": 1759, "95th": 2660, "99th": 7069, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00233-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00250-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b96eb1b7fe7094b866c8e1e944ccd86898f45993dc37dd71b7ab3e428233197e
3
+ size 25966910
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00266-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98e3a766fce178b6992bbee190d6fe7981822ef04ce4f60854038962b398626a
3
+ size 67106424
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00266-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3eb8b5114fce3681bc25df5f05f1db2b07b61a0691c5ac270d60c2f2b85a58ec
3
+ size 67103844
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9f69e8e29fdf7767b99e14d53116a5be8fba1753d213c08c901b42ce9af71d2
3
+ size 67106940
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d830faa5d124bf1d6a717b3e11e6a51ffbe8eb769d81fb2ac33645f0f4f12b86
3
+ size 67106263
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00341-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67077295, "hashes": {}}, "samples": 18588, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 25419563, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17583671, "hashes": {}}, "samples": 5047, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6726237, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00341-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 20788084, "total_tokens_skipped": 6, "percentiles": {"0th": 59, "10th": 196, "20th": 264, "30th": 325, "40th": 386, "50th": 481, "60th": 631, "70th": 816, "80th": 1136, "90th": 1875, "95th": 2942, "99th": 6935, "100th": 8191}}
train/fineweb2-sampled-decay-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00341-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff