Datasets:

ArXiv:
License:
orionweller commited on
Commit
71463f7
·
verified ·
1 Parent(s): 97bd0a1

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0102-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  2. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0147-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  3. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0147-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  4. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0147-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  5. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  6. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0023-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  7. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0133-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  8. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0175-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  9. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0175-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  10. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0175-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  11. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0022-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  12. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  13. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  14. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  15. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  16. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0069-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  17. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  18. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  19. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  20. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/004_00000-batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  21. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/004_00000-batch_0114-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  22. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
  23. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00008-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  24. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  25. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  26. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  27. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  28. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  29. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  30. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  31. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  32. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  33. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  34. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
  36. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  37. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  38. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  39. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  40. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  41. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  42. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  43. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  44. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  45. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  46. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  47. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  48. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  49. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  50. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0102-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0147-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106497, "hashes": {}}, "samples": 14922, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14315975, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 23379260, "hashes": {}}, "samples": 5465, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5222798, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0147-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 22296264, "total_tokens_skipped": 191, "percentiles": {"0th": 72, "10th": 165, "20th": 239, "30th": 329, "40th": 446, "50th": 603, "60th": 816, "70th": 1098, "80th": 1527, "90th": 2461, "95th": 3650, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0147-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a202b8bbbea33ce975a0c184d453de2ae992753d3644f28b5a468764432b967c
3
+ size 67108446
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0023-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f70d919f7f26fc7f8ebc4c933720c9642e65411775f46f092c58a0c83d05a52
3
+ size 67106988
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0133-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5073e5c67402ddda01d379eafe694a2885f10a405de25d5575cf6eb19b1f8618
3
+ size 523606
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0175-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67100876, "hashes": {}}, "samples": 18332, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6603658, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106419, "hashes": {}}, "samples": 20868, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6495220, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108274, "hashes": {}}, "samples": 20530, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6710829, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 2163491, "hashes": {}}, "samples": 665, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 223348, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0175-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 49906260, "total_tokens_skipped": 0, "percentiles": {"0th": 91, "10th": 190, "20th": 255, "30th": 320, "40th": 399, "50th": 490, "60th": 616, "70th": 788, "80th": 1060, "90th": 1669, "95th": 2517, "99th": 7157, "100th": 8191}}
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0175-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0022-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecc61897fec49604f97c621eb3bb6f927f3ddfd56b036db4c89bcf0e74810d06
3
+ size 67107369
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/001_00000-batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1a978ad60b205388db09bdb68fcf5c85667932b553aa0296f11e5e63f13a07a
3
+ size 9554427
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106798, "hashes": {}}, "samples": 20863, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8654421, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105581, "hashes": {}}, "samples": 21063, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9056510, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 55985315, "hashes": {}}, "samples": 16033, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8159458, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 46624559, "total_tokens_skipped": 120, "percentiles": {"0th": 90, "10th": 200, "20th": 268, "30th": 339, "40th": 415, "50th": 512, "60th": 640, "70th": 817, "80th": 1096, "90th": 1669, "95th": 2354, "99th": 4896, "100th": 8191}}
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0069-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73fa1891e52e5d0085a64402ea1fc281bd0cb2c0bdfe33356d57d7a09fc1725c
3
+ size 63117409
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108675, "hashes": {}}, "samples": 19256, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13078074, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107531, "hashes": {}}, "samples": 22283, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11628366, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 5079761, "hashes": {}}, "samples": 1441, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 862117, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 34138195, "total_tokens_skipped": 8, "percentiles": {"0th": 92, "10th": 195, "20th": 262, "30th": 327, "40th": 401, "50th": 490, "60th": 609, "70th": 772, "80th": 1022, "90th": 1597, "95th": 2351, "99th": 5784, "100th": 8191}}
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/004_00000-batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e53590a742fcb0e77873c24610f2a715b3c7cb027d2396ec4b03f84b1689a55e
3
+ size 67095899
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/004_00000-batch_0114-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c078d1427daa2d20b0ace027a09a8abf5a72247a91ed4e1e1528c8d8fc6a10
3
+ size 67107564
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:542d95cf30865bb368764f52f6e7686057aa46831eabe082d9958c8b8a3dd8eb
3
+ size 53942590
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00008-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9166e9a3dcacb89ad702d960fbdc9110506f260f4ca64f6d677fc468de06f2da
3
+ size 15048566
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc0b16c03fbefe712b83e917e7b3a5f9411aecc120dd0696224037b5083ae692
3
+ size 67084528
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:147ef23e89c49e9b58d55ae91b903b7226aa21f95c7da5f331e2f8d410dae553
3
+ size 67107827
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a509c26d8d7bd6937ba99898be6f0eb08878c3d6b59e9bc33a7fc8c711025d86
3
+ size 67083771
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67085791, "hashes": {}}, "samples": 13432, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11598823, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108222, "hashes": {}}, "samples": 13036, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11776297, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67095604, "hashes": {}}, "samples": 12797, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 12549341, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 39981182, "hashes": {}}, "samples": 7956, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6953537, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 59564554, "total_tokens_skipped": 78, "percentiles": {"0th": 80, "10th": 158, "20th": 235, "30th": 335, "40th": 420, "50th": 525, "60th": 721, "70th": 1033, "80th": 1630, "90th": 3242, "95th": 6106, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e358e6eebc0e1de02af3f00cab8d94ffb56a2a2ee835f08cb04e75b7a020a83
3
+ size 67104331
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c90efa5ce2c8577173818dfe6013a682344fbdfc9e892db820ec44b6acc79ca
3
+ size 67093737
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6b8f5b0a28977313e7de143df041baad865a7c5895df0ef5335d5898157b3a5
3
+ size 67101149
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ed5a50cd26852ae1127e295cad5f20d12e1158eef0ec9df3f2c15b2aa32777a
3
+ size 50650353
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106890, "hashes": {}}, "samples": 13729, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8022400, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107311, "hashes": {}}, "samples": 12469, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8559603, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67099961, "hashes": {}}, "samples": 18029, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6562436, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107166, "hashes": {}}, "samples": 14536, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7787219, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67105690, "hashes": {}}, "samples": 13233, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 8727328, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 14258052, "hashes": {}}, "samples": 1792, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 1608128, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad896727bf02d92eeed0e94e2601c8502e6677b98dd2ec82a1540486df123698
3
+ size 14258052
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 86269690, "total_tokens_skipped": 160, "percentiles": {"0th": 87, "10th": 153, "20th": 221, "30th": 318, "40th": 404, "50th": 497, "60th": 668, "70th": 953, "80th": 1476, "90th": 2873, "95th": 5305, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43cbf89e55ecc21ce1d3ddfa2c8a773c4537d3c1de8abe7220a82c7fe0a6065a
3
+ size 67107431
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8f5bd7c7c8f18eb4d98814a9fbcc4570ed81ca10164e8a7ce14ebb5fe783459
3
+ size 67076562
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:129600da5c6de84eae34358cc8f4b06f45d84da6cc62b08dbda65b65d09119f0
3
+ size 67107396
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67077798, "hashes": {}}, "samples": 15097, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9286057, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107250, "hashes": {}}, "samples": 13956, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10404196, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107617, "hashes": {}}, "samples": 14849, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9133941, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108844, "hashes": {}}, "samples": 15721, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 9649821, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 4633202, "hashes": {}}, "samples": 994, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 711460, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cdda9b931b82f2904195d9ac98aba7d99e371a3c2b2f3d0be377078e741340d
3
+ size 4633202
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 67292199, "total_tokens_skipped": 102, "percentiles": {"0th": 81, "10th": 153, "20th": 219, "30th": 313, "40th": 395, "50th": 478, "60th": 639, "70th": 919, "80th": 1416, "90th": 2721, "95th": 4808, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55281ed5e47398b31f60e5c13e20f76b05e364835604fe0c6a0459ae8253849b
3
+ size 25530033
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d975f816bc2c364690e4c0b1f83175f7c124a2cc0270700dfa3f14ede5378417
3
+ size 67104787
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8cb9a5091e274440d321152ee23b6e64871f288b8db5291cf74f1ce9d85c5f8
3
+ size 67100394
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107975, "hashes": {}}, "samples": 11150, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 19247748, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108051, "hashes": {}}, "samples": 11318, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 19131111, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 23691699, "hashes": {}}, "samples": 3748, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6691626, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 39058598, "total_tokens_skipped": 96, "percentiles": {"0th": 86, "10th": 163, "20th": 261, "30th": 373, "40th": 475, "50th": 645, "60th": 916, "70th": 1304, "80th": 2091, "90th": 4374, "95th": 7027, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff