Datasets:

ArXiv:
License:
orionweller commited on
Commit
d4bb57b
·
verified ·
1 Parent(s): 7a497bc

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  2. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  3. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  4. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  5. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0082-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  6. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  7. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  8. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  9. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0146-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  10. train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  11. train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  12. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  13. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  14. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  15. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  16. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  17. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00130-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  18. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00130-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  19. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00130-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  20. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00151-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  21. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00160-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  22. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00166-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  23. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00274-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  24. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00285-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  25. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00576-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  26. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00577-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  27. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00657-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  28. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00674-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  29. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00687-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  30. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00687-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  31. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00749-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  32. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00749-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  33. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00749-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  34. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00906-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  35. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00930-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  36. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00963-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  37. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00049-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  38. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00095-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  39. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00126-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  40. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00127-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  41. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00166-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  42. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00166-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  43. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  44. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  45. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  46. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00222-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  47. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00222-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  48. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00222-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  49. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00325-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  50. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00325-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67088658, "hashes": {}}, "samples": 16214, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8695735, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17491258, "hashes": {}}, "samples": 4046, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2518159, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 20821813, "total_tokens_skipped": 55, "percentiles": {"0th": 70, "10th": 163, "20th": 237, "30th": 330, "40th": 440, "50th": 571, "60th": 744, "70th": 1007, "80th": 1412, "90th": 2299, "95th": 3439, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0082-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b34fbbfcc2090421ed013d63764f519fe44e710acd95d157a96ee5a4e41eac86
3
+ size 11897572
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67097590, "hashes": {}}, "samples": 15604, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11573652, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18310283, "hashes": {}}, "samples": 4741, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3306510, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21027491, "total_tokens_skipped": 0, "percentiles": {"0th": 62, "10th": 160, "20th": 233, "30th": 310, "40th": 408, "50th": 556, "60th": 731, "70th": 983, "80th": 1393, "90th": 2323, "95th": 3610, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0146-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f5b157714088bb5c3c685ed3a1599dc2a06cb7476a7ff7b68de4ec198948475
3
+ size 21855180
train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108665, "hashes": {}}, "samples": 11084, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5634602, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108660, "hashes": {}}, "samples": 10787, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5504444, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107393, "hashes": {}}, "samples": 11643, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5619528, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67102632, "hashes": {}}, "samples": 10481, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5603328, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67107747, "hashes": {}}, "samples": 10170, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5772672, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 67106797, "hashes": {}}, "samples": 10408, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 5452921, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00006.mds", "bytes": 37716053, "hashes": {}}, "samples": 5048, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00006.mds.zstd", "bytes": 3401564, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67098253, "hashes": {}}, "samples": 17616, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7499443, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105333, "hashes": {}}, "samples": 16738, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7568972, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67094341, "hashes": {}}, "samples": 18545, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7748847, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67104297, "hashes": {}}, "samples": 17764, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7822632, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 25324238, "hashes": {}}, "samples": 7354, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 3254797, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 72187976, "total_tokens_skipped": 209, "percentiles": {"0th": 61, "10th": 169, "20th": 226, "30th": 288, "40th": 354, "50th": 445, "60th": 621, "70th": 857, "80th": 1236, "90th": 2098, "95th": 3303, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107705, "hashes": {}}, "samples": 19343, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6979999, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108432, "hashes": {}}, "samples": 17770, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7450215, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67102721, "hashes": {}}, "samples": 18063, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7203643, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67096881, "hashes": {}}, "samples": 16479, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7426685, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 56355495, "hashes": {}}, "samples": 14145, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 6962799, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00130-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33051230d6fe19187c1177ed95b60c9b42bc905c810424e96612c72475ce5652
3
+ size 67107416
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00130-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a8338157d224b4d27dc1713236378e69aaa40de6038ca14669c1a6b980c2a51
3
+ size 67105482
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00130-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dbbc331a411c1464cf1417d03ffe86b3b77309c8b37c4a477130ba490352f41
3
+ size 67106947
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00151-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8176884a8c670a24524e81ed342c04cb5f54787d66a663dd260b5892f5f1b634
3
+ size 3230563
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00160-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cf9305ff0a1cdc2677fc77964358b99883fdd8f5badb297584f55e86ed03f66
3
+ size 42183625
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00166-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:878133048517d4c9c45c351a63e8118fb952fd9357c23459917549f4af251617
3
+ size 67080484
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00274-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b66791c608b1da497b8e6d168401868862c1fdd941c3ce0d906ee1cb74c1091
3
+ size 20599942
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00285-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:564ff2d328f10eb815dbcd7dad21374c27555cff8968da159435313eb45739b6
3
+ size 66436586
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00576-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:960e353b71c870f3a6205df245658e0be41237aeae0868b2d4106ee1d6d0d02a
3
+ size 67107215
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00577-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75b58af3ee708092eaa6e038cf2801da370b8ee051bfb0e99eea868b9f3c63fd
3
+ size 67106654
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00657-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a71126f31f319f2943a444e84b2d50c7e4f384aaad9dc507220e7161fe240ef
3
+ size 43277166
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00674-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:147cecf857fbd677d76fe03e597c247ae2dd62080384291895bdb39e212301c9
3
+ size 67107109
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00687-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8ff9e694b379efcd11731072d007143c745161faeda117909a7775c14dfba17
3
+ size 67108417
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00687-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f9b0b203e788f53f8bd2cbd7bdc9063f5f2a2a5b5020a5c114a7b0de884ad96
3
+ size 67094424
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00749-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104090, "hashes": {}}, "samples": 16165, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11788420, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67098912, "hashes": {}}, "samples": 16586, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12376509, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67099395, "hashes": {}}, "samples": 16067, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 13376788, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 8935477, "hashes": {}}, "samples": 2043, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 1749726, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00749-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 51747840, "total_tokens_skipped": 145, "percentiles": {"0th": 54, "10th": 188, "20th": 266, "30th": 335, "40th": 414, "50th": 560, "60th": 739, "70th": 983, "80th": 1423, "90th": 2335, "95th": 3433, "99th": 8097, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00749-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00906-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00930-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d13897fab8e0599536a9cb460bdf7c5e50a32217a36a5150cfc85a80ce47c4d
3
+ size 54186292
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00963-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fa22fa687f49fce6115e5ca75699eb6c81886f2c4cf7ef822bc089ca049253d
3
+ size 33086493
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00049-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8279eb36409964cd1a2f16c6a9fa1feaec82bdfd097befdb0c75d0bdad8af75a
3
+ size 66123659
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00095-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b0d1ad742437c67081a553dd94619047f4b1b81f15193cb43e5f9e25707bb82
3
+ size 67100128
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00126-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00127-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95034294045ee26be8ee3438c9514f8a9164de3540ebe8d2bbce553419354c66
3
+ size 67108299
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00166-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4ca301e9b4a62147b0e34603f42918135de7b00040dbbb16495ebb7c075427b
3
+ size 67093785
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00166-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f7885d2456a8b227557545e7480a8aa371405dbae9b400527f317afc17ff38f
3
+ size 67093870
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67088798, "hashes": {}}, "samples": 21473, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8301108, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106753, "hashes": {}}, "samples": 20557, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7613546, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108362, "hashes": {}}, "samples": 21243, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8530656, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 43348454, "hashes": {}}, "samples": 13714, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 4967407, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 59938045, "total_tokens_skipped": 44, "percentiles": {"0th": 61, "10th": 108, "20th": 145, "30th": 209, "40th": 282, "50th": 353, "60th": 470, "70th": 679, "80th": 985, "90th": 1759, "95th": 2922, "99th": 8189, "100th": 8191}}
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00222-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:658b9331071b302af23f5afeebf81f5be9ce0c61a3095e11dcf3ea2215600d6d
3
+ size 67107621
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00222-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:968f55b76a8aa29b6c5a723f8698d37a863813709a647f2f5a243032a4be3ffa
3
+ size 67107782
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00222-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93c1d19f95d78148e3a1a719df8bab167a30b76bfcca85305ec4b3b1dd98ad85
3
+ size 67107295
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00325-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d44ad518ccea69133165f0d861e1df10345cec673d349622554816df96841e1
3
+ size 67108383
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00325-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c951cd049ccbfae288cc7176411e9e7c4a1a2f9cc53d6f7adc58bbb93e3b258
3
+ size 67107247