Datasets:

ArXiv:
License:
orionweller commited on
Commit
a7770db
·
verified ·
1 Parent(s): 9fb5a55

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  2. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  3. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  4. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  5. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  6. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  7. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  8. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  9. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0053-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  10. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0053-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  11. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0053-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  12. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0055-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  13. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0055-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  14. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0055-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  15. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  16. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  17. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  18. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0092-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  19. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0092-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  20. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0092-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  21. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0104-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  22. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0104-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  23. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0104-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  24. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  25. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0131-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  26. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0142-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  27. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  28. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  29. train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  30. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  31. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  32. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00067-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  33. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00193-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  34. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00239-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  35. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00252-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  36. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00252-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  37. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00252-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  38. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  39. train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00438-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  40. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  41. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  42. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0049-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  43. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0050-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  44. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0050-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  45. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0083-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  46. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0083-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  47. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  48. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0122-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  49. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0156-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  50. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0160-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103901, "hashes": {}}, "samples": 14472, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8416717, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 35964830, "hashes": {}}, "samples": 6126, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3463597, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 25438587, "total_tokens_skipped": 0, "percentiles": {"0th": 74, "10th": 166, "20th": 246, "30th": 337, "40th": 460, "50th": 632, "60th": 847, "70th": 1143, "80th": 1597, "90th": 2882, "95th": 5158, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67097938, "hashes": {}}, "samples": 16246, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8399717, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22081584, "hashes": {}}, "samples": 4143, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2390373, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21969690, "total_tokens_skipped": 22, "percentiles": {"0th": 70, "10th": 161, "20th": 234, "30th": 317, "40th": 436, "50th": 579, "60th": 772, "70th": 1044, "80th": 1465, "90th": 2412, "95th": 3823, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106818, "hashes": {}}, "samples": 13656, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8284415, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 28286518, "hashes": {}}, "samples": 6796, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4127575, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 23522094, "total_tokens_skipped": 0, "percentiles": {"0th": 75, "10th": 165, "20th": 242, "30th": 342, "40th": 459, "50th": 613, "60th": 806, "70th": 1075, "80th": 1549, "90th": 2599, "95th": 4409, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0053-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67083361, "hashes": {}}, "samples": 14780, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9354571, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 24170950, "hashes": {}}, "samples": 5580, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3399074, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0053-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 22488796, "total_tokens_skipped": 21, "percentiles": {"0th": 73, "10th": 163, "20th": 244, "30th": 334, "40th": 460, "50th": 608, "60th": 806, "70th": 1073, "80th": 1529, "90th": 2494, "95th": 3860, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0053-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0055-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101667, "hashes": {}}, "samples": 16953, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9303906, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17924572, "hashes": {}}, "samples": 3286, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2711956, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0055-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 20933768, "total_tokens_skipped": 150, "percentiles": {"0th": 68, "10th": 162, "20th": 236, "30th": 318, "40th": 432, "50th": 573, "60th": 765, "70th": 1033, "80th": 1450, "90th": 2345, "95th": 3561, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0055-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108540, "hashes": {}}, "samples": 16092, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10579963, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20861934, "hashes": {}}, "samples": 4298, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3411940, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21667397, "total_tokens_skipped": 0, "percentiles": {"0th": 66, "10th": 158, "20th": 238, "30th": 322, "40th": 436, "50th": 581, "60th": 778, "70th": 1046, "80th": 1457, "90th": 2364, "95th": 3611, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0092-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105983, "hashes": {}}, "samples": 15353, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10656836, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21190313, "hashes": {}}, "samples": 5026, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3390859, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0092-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21749029, "total_tokens_skipped": 27, "percentiles": {"0th": 72, "10th": 161, "20th": 238, "30th": 325, "40th": 442, "50th": 584, "60th": 785, "70th": 1034, "80th": 1444, "90th": 2373, "95th": 3634, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0092-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0104-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105212, "hashes": {}}, "samples": 16633, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11654202, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15174584, "hashes": {}}, "samples": 3598, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2868044, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0104-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 20247299, "total_tokens_skipped": 0, "percentiles": {"0th": 68, "10th": 162, "20th": 234, "30th": 315, "40th": 422, "50th": 558, "60th": 741, "70th": 994, "80th": 1407, "90th": 2281, "95th": 3313, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0104-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0108-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e41fa8c9a990d6e36361982d8561c79442166959cebee70b96c30cf76960cfea
3
+ size 18310283
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0131-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a18eeb08e4b5f645ec6efa810c0e43ca9610f17abcde9630b8d4045294484004
3
+ size 16449994
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0142-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108444, "hashes": {}}, "samples": 15752, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14301882, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19937837, "hashes": {}}, "samples": 4546, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4195370, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67095199, "hashes": {}}, "samples": 12692, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 21195233, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 38783805, "hashes": {}}, "samples": 7815, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12343975, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 26142566, "total_tokens_skipped": 37, "percentiles": {"0th": 66, "10th": 166, "20th": 254, "30th": 369, "40th": 518, "50th": 710, "60th": 951, "70th": 1299, "80th": 1926, "90th": 2817, "95th": 4429, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/afr_Latn_train-sampled/batch_0187-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e566db91443e03b665844c1741374bcff61588080593974675b4bf4ce6f9cc57
3
+ size 67107147
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f4ad8ad8dbd3769c447a488176bee63bd5d57bf38cef8fbaf1175931ee7d31f
3
+ size 67108709
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00067-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e51be47c56be557ab69fd9c8c1cfc46d19394fc007bea0cd081bc5e87237de50
3
+ size 36983898
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00193-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9d2ca1e43c4f1c3450b6559168d890fba84ca71ea3583032c423bbdfe2c8ade
3
+ size 55590806
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00239-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bb6696bbb7ab99df7f8d6292dce1a5ffd06059a935e4051445d7467b1d3a1fa
3
+ size 67108332
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00252-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107452, "hashes": {}}, "samples": 19683, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11456311, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108824, "hashes": {}}, "samples": 18739, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9906593, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 30352882, "hashes": {}}, "samples": 6749, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5194062, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00252-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 40422816, "total_tokens_skipped": 102, "percentiles": {"0th": 63, "10th": 108, "20th": 159, "30th": 253, "40th": 335, "50th": 444, "60th": 625, "70th": 858, "80th": 1214, "90th": 1985, "95th": 3258, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00252-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0f9c25bb18c6380139c02abdf2f3ac9ec3f91c312a927652f8cf1f04e866615
3
+ size 62053495
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00438-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f9bd09ed67b2928024a7e1a3152ade5a1c779c2bd220416927b4ecb5848f536
3
+ size 67107995
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab0c08b0d0b1f768db461a91dddd11f60ac25cbb3f9fb34c80133c3f27d3fdcf
3
+ size 67107978
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0027-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abaadc4c51239dfb7aa7ba68f4ac523951cdfef2c5c7e6067883ad4b75fb2df7
3
+ size 67106204
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0049-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfc36dfe444cb2a6d450c4152c3899afe4883a54de97bfc58eca490d698339f8
3
+ size 67103105
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0050-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd59041a9afb907c08134774b375220b066277576feda35d9cc10539e635b5be
3
+ size 67107227
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0050-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b1346ec1fc937ce6e3a7595618ff80d731b8c3f6f7bb3675fa085671b3b812f
3
+ size 62977386
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0083-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5a7131f27a80e9947fc16a3f7cd998ebb27ffe158ee6b07724770fbd23ed678
3
+ size 67105639
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0083-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30bde2915b67ffeec96848692f8d7836dd5b3bdae244933e42d850a82451013a
3
+ size 63590907
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2f6b75bfcae2ee8addc965f3df108fac7da36658c8690114bdd93612bbb7b7e
3
+ size 67107749
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0122-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7daccc8dcfb0c27382be852ef2184948d702d45a9ccbd4e7cd0d4dd168d3a4a6
3
+ size 57149092
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0156-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90a9fa801af9372c31ffb1bd0964dd05e76524c7f9bb222852562a47302dbffb
3
+ size 67107582
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0160-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99aeb1d3570f1ee76b82ab01a6ed601f0108073e2babcda83ac7712a95838c81
3
+ size 67087897