Datasets:

ArXiv:
License:
orionweller commited on
Commit
6bf143f
·
verified ·
1 Parent(s): b1770df

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  2. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  3. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  4. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_35-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  5. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_35-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  6. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_35-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  7. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  8. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  9. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  10. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json +1 -0
  11. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json +3 -0
  12. train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json +12 -0
  13. train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  14. train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  15. train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  16. train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  17. train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  18. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  19. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  20. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  21. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  22. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  23. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  24. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  25. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  26. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  27. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json +1 -0
  28. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json +3 -0
  29. train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json +12 -0
  30. train/multi-wikis-sampled-decay/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  31. train/multi-wikis-sampled-decay/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  32. train/multi-wikis-sampled-decay/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  33. train/multi-wikis-sampled-decay/kor_Hang-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  34. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  36. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  37. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  38. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  39. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  40. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json +1 -0
  41. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json +3 -0
  42. train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json +11 -0
  43. train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  44. train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json +1 -0
  45. train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json +3 -0
  46. train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json +12 -0
  47. train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_26-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  48. train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_26-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  49. train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_26-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  50. train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_32-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 24615022, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9377150, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 5802626, "total_tokens_skipped": 0, "percentiles": {"0th": 20, "10th": 42, "20th": 50, "30th": 58, "40th": 67, "50th": 79, "60th": 94, "70th": 117, "80th": 155, "90th": 236, "95th": 326, "99th": 558, "100th": 3624}}
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_35-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 24506037, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9253928, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_35-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 5750409, "total_tokens_skipped": 0, "percentiles": {"0th": 20, "10th": 42, "20th": 50, "30th": 58, "40th": 67, "50th": 78, "60th": 93, "70th": 117, "80th": 153, "90th": 231, "95th": 325, "99th": 551, "100th": 3343}}
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_35-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 25146778, "hashes": {}}, "samples": 50020, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9408337, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 5922989, "total_tokens_skipped": 0, "percentiles": {"0th": 20, "10th": 42, "20th": 50, "30th": 58, "40th": 66, "50th": 77, "60th": 92, "70th": 114, "80th": 151, "90th": 228, "95th": 322, "99th": 564, "100th": 8191}}
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"version": 2, "shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_8-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 25146778, "hashes": {}}, "samples": 50020, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_8-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9408337, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_35-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 24506037, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_35-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9253928, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 24615022, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9377150, "hashes": {}}}]}
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "num_tokens": 17476024
3
+ }
train/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_tokens": 17476024,
3
+ "target_tokens": 15000000,
4
+ "num_unique_folders": 3,
5
+ "num_total_folders_copied": 3,
6
+ "num_upsampled_folders": 0,
7
+ "copied_folders": [
8
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups",
9
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups",
10
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_35-tokenized-chunked-8192-512-32-backfill-nodups"
11
+ ]
12
+ }
train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 48920622, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16551306, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 11876221, "total_tokens_skipped": 0, "percentiles": {"0th": 22, "10th": 80, "20th": 102, "30th": 126, "40th": 150, "50th": 179, "60th": 214, "70th": 261, "80th": 333, "90th": 463, "95th": 601, "99th": 969, "100th": 4564}}
train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 11986258, "total_tokens_skipped": 0, "percentiles": {"0th": 34, "10th": 81, "20th": 103, "30th": 127, "40th": 152, "50th": 180, "60th": 216, "70th": 263, "80th": 336, "90th": 465, "95th": 602, "99th": 986, "100th": 8190}}
train/multi-wikis-sampled-decay/heb_Hebr-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 26290837, "hashes": {}}, "samples": 50033, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7917230, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 6208964, "total_tokens_skipped": 0, "percentiles": {"0th": 18, "10th": 41, "20th": 55, "30th": 68, "40th": 76, "50th": 85, "60th": 91, "70th": 111, "80th": 147, "90th": 223, "95th": 312, "99th": 586, "100th": 8191}}
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 25420544, "hashes": {}}, "samples": 50018, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7931620, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 5991482, "total_tokens_skipped": 0, "percentiles": {"0th": 18, "10th": 41, "20th": 54, "30th": 68, "40th": 75, "50th": 80, "60th": 91, "70th": 111, "80th": 149, "90th": 225, "95th": 311, "99th": 561, "100th": 8191}}
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 24996252, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7961420, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 5885465, "total_tokens_skipped": 0, "percentiles": {"0th": 19, "10th": 40, "20th": 54, "30th": 67, "40th": 74, "50th": 79, "60th": 90, "70th": 112, "80th": 152, "90th": 234, "95th": 323, "99th": 578, "100th": 8190}}
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"version": 2, "shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_5-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 24996252, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_5-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 7961420, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_3-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 25420544, "hashes": {}}, "samples": 50018, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_3-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 7931620, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_10-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 26290837, "hashes": {}}, "samples": 50033, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_10-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 7917230, "hashes": {}}}]}
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "num_tokens": 18085911
3
+ }
train/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_tokens": 18085911,
3
+ "target_tokens": 15000000,
4
+ "num_unique_folders": 3,
5
+ "num_total_folders_copied": 3,
6
+ "num_upsampled_folders": 0,
7
+ "copied_folders": [
8
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups",
9
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_10-tokenized-chunked-8192-512-32-backfill-nodups",
10
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/ind_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_3-tokenized-chunked-8192-512-32-backfill-nodups"
11
+ ]
12
+ }
train/multi-wikis-sampled-decay/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ade1c04643dbd8af9aaa5da61b1e1c81ae4583925f05e86a9cc0250c3c0e8e0f
3
+ size 23070124
train/multi-wikis-sampled-decay/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:178cb83a97e6932d53f4cb827d808a49f4129951b1eb1aaa17a9b9a2b3bfb9c7
3
+ size 26253464
train/multi-wikis-sampled-decay/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf2ce8bd378f0a56873f0aafe371c320d158e6b012ad0b9ac386038346542d25
3
+ size 22464872
train/multi-wikis-sampled-decay/kor_Hang-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:344f925348274329e2de0c22e2083066952cb41585a546daf8f7541861c4b577
3
+ size 29601561
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 32838919, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10070680, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 7860636, "total_tokens_skipped": 0, "percentiles": {"0th": 28, "10th": 56, "20th": 67, "30th": 79, "40th": 94, "50th": 112, "60th": 133, "70th": 166, "80th": 214, "90th": 308, "95th": 414, "99th": 723, "100th": 8190}}
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 33713568, "hashes": {}}, "samples": 50004, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10334027, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 8076364, "total_tokens_skipped": 0, "percentiles": {"0th": 30, "10th": 55, "20th": 66, "30th": 79, "40th": 94, "50th": 111, "60th": 135, "70th": 169, "80th": 221, "90th": 319, "95th": 432, "99th": 749, "100th": 8191}}
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"version": 2, "shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 32838919, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 10070680, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 33713568, "hashes": {}}, "samples": 50004, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 10334027, "hashes": {}}}]}
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "num_tokens": 15937000
3
+ }
train/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_tokens": 15937000,
3
+ "target_tokens": 15000000,
4
+ "num_unique_folders": 2,
5
+ "num_total_folders_copied": 2,
6
+ "num_upsampled_folders": 0,
7
+ "copied_folders": [
8
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups",
9
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/lit_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups"
10
+ ]
11
+ }
train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 40158581, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12055670, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"version": 2, "shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 40158581, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 12055670, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_2-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 9114273, "hashes": {}}, "samples": 11993, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_2-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 2661329, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 39259621, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 11691544, "hashes": {}}}]}
train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/num_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "num_tokens": 21342588
3
+ }
train/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/sampling_summary.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_tokens": 21342588,
3
+ "target_tokens": 15000000,
4
+ "num_unique_folders": 3,
5
+ "num_total_folders_copied": 3,
6
+ "num_upsampled_folders": 0,
7
+ "copied_folders": [
8
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_0-tokenized-chunked-8192-512-32-backfill-nodups",
9
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_2-tokenized-chunked-8192-512-32-backfill-nodups",
10
+ "/home/oweller2/my_scratch/bert24-data/data/mmbert-ext/multi-wikis-sampled-decay/lvs_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_1-tokenized-chunked-8192-512-32-backfill-nodups"
11
+ ]
12
+ }
train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_26-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 22237538, "hashes": {}}, "samples": 50003, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6405018, "hashes": {}}}], "version": 2}
train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_26-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 5183449, "total_tokens_skipped": 0, "percentiles": {"0th": 18, "10th": 47, "20th": 53, "30th": 56, "40th": 59, "50th": 64, "60th": 74, "70th": 92, "80th": 125, "90th": 206, "95th": 300, "99th": 526, "100th": 8191}}
train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_26-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/multi-wikis-sampled-decay/nld_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/articles_32-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 22236291, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6381171, "hashes": {}}}], "version": 2}