Datasets:

ArXiv:
License:
orionweller commited on
Commit
f13dbcd
·
verified ·
1 Parent(s): a7770db

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00249-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  2. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00249-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  3. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00282-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  4. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00863-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  5. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00863-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  6. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  7. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  8. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  9. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  10. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  11. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  12. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00077-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  13. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00104-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  14. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  15. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  16. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  17. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  18. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  19. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  20. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00232-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  21. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00251-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  22. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00253-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  23. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00253-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  24. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00253-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  25. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00260-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  26. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00260-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  27. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00269-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  28. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  29. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00308-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  30. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00308-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  31. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00308-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  32. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00347-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  33. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00347-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  34. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00347-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  35. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00348-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  36. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00351-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  37. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00351-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  38. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  39. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  40. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  41. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00399-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  42. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00399-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  43. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00399-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  44. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00421-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  45. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00425-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  46. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0184-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  47. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  48. train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00012-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  49. train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  50. train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00249-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 58244959, "hashes": {}}, "samples": 16324, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15622626, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00249-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 14300783, "total_tokens_skipped": 0, "percentiles": {"0th": 65, "10th": 196, "20th": 262, "30th": 326, "40th": 391, "50th": 481, "60th": 639, "70th": 845, "80th": 1189, "90th": 1928, "95th": 2804, "99th": 6321, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00282-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 13863164, "total_tokens_skipped": 6, "percentiles": {"0th": 58, "10th": 192, "20th": 265, "30th": 325, "40th": 387, "50th": 476, "60th": 626, "70th": 815, "80th": 1147, "90th": 1849, "95th": 2858, "99th": 7092, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00863-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106388, "hashes": {}}, "samples": 17554, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 22037659, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 56326828, "hashes": {}}, "samples": 15708, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12912161, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00863-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:519c0da4a4f6bf01372828c735ee8965e903c036cb3c8b9e789fdf201c7e522b
3
+ size 67086011
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c56e12ec3c04faac8525ce28e4aaa0da7e8e9c9667fca27e500eff20ccc35b2
3
+ size 67100745
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f5fcbca79314700d124000e61ff53885c3bac9953594425daefdcf76d779b59
3
+ size 67108647
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103805, "hashes": {}}, "samples": 18766, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9362862, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106374, "hashes": {}}, "samples": 20171, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9759675, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 64416884, "hashes": {}}, "samples": 19565, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 10127126, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 48725510, "total_tokens_skipped": 50, "percentiles": {"0th": 60, "10th": 102, "20th": 149, "30th": 228, "40th": 302, "50th": 387, "60th": 545, "70th": 738, "80th": 1027, "90th": 1793, "95th": 3129, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00077-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b062a065c0d28d1ce66441fd1e99a2c891b016ae25edf79c42f35a7dfb2d957
3
+ size 38990191
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00104-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32bb2ed8b097951d1f4eeaf712bfcb6b523c8e820074e3cc585ced9467122ecc
3
+ size 11277611
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106777, "hashes": {}}, "samples": 17212, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11118528, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105391, "hashes": {}}, "samples": 16994, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12055058, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 66782315, "hashes": {}}, "samples": 16730, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 12071145, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 49437381, "total_tokens_skipped": 11, "percentiles": {"0th": 60, "10th": 105, "20th": 161, "30th": 249, "40th": 332, "50th": 455, "60th": 649, "70th": 909, "80th": 1304, "90th": 2178, "95th": 3734, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103128, "hashes": {}}, "samples": 17821, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 19965075, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 35120636, "hashes": {}}, "samples": 9384, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11173740, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 25122514, "total_tokens_skipped": 0, "percentiles": {"0th": 58, "10th": 106, "20th": 173, "30th": 265, "40th": 351, "50th": 499, "60th": 682, "70th": 920, "80th": 1280, "90th": 2030, "95th": 3165, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00232-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e0dcc0e678eb518308a83b4ca7a3ddf78e1f06ea1625b9695a9f9f37fa3bf02
3
+ size 9494270
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00251-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3689efe8de39b0aa02d1dc63063a21e2cfd499dc059410e2fe06a4d3eca92618
3
+ size 67105562
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00253-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108781, "hashes": {}}, "samples": 18033, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6965236, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108311, "hashes": {}}, "samples": 21595, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7322443, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67095007, "hashes": {}}, "samples": 21580, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7318678, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67106743, "hashes": {}}, "samples": 18940, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7216458, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 8674265, "hashes": {}}, "samples": 2611, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 976272, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00253-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 67955795, "total_tokens_skipped": 140, "percentiles": {"0th": 59, "10th": 102, "20th": 150, "30th": 229, "40th": 303, "50th": 391, "60th": 545, "70th": 747, "80th": 1057, "90th": 1818, "95th": 3000, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00253-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00260-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67091846, "hashes": {}}, "samples": 19863, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7264327, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108666, "hashes": {}}, "samples": 20750, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6919730, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67103623, "hashes": {}}, "samples": 20139, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6762074, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107514, "hashes": {}}, "samples": 20623, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6903686, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 21820029, "hashes": {}}, "samples": 4368, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 2381865, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00260-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00269-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6c713613589e80a0fd51b1d841c3ca19cb513fe500cf41133c8698c7d5c3dc5
3
+ size 67106566
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00270-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:714987de5bdb81eb61cb8407f9c05420188fb1aebe0e63b8edfd110d0da39e1e
3
+ size 67108164
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00308-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106266, "hashes": {}}, "samples": 15694, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15223253, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18429349, "hashes": {}}, "samples": 4563, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5076325, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00308-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21061189, "total_tokens_skipped": 11, "percentiles": {"0th": 62, "10th": 105, "20th": 168, "30th": 258, "40th": 345, "50th": 471, "60th": 653, "70th": 920, "80th": 1355, "90th": 2432, "95th": 4432, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00308-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00347-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104928, "hashes": {}}, "samples": 17744, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 22717637, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 23278551, "hashes": {}}, "samples": 6103, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8379728, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00347-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 22215915, "total_tokens_skipped": 0, "percentiles": {"0th": 60, "10th": 107, "20th": 175, "30th": 267, "40th": 348, "50th": 489, "60th": 674, "70th": 916, "80th": 1286, "90th": 2082, "95th": 3247, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00347-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00348-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cb702d8af348cc25f4bc3ba2686434c00595f0e17f5eaa66f682d2d5893b2a9
3
+ size 67104484
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00351-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 45991750, "total_tokens_skipped": 0, "percentiles": {"0th": 59, "10th": 99, "20th": 144, "30th": 220, "40th": 296, "50th": 385, "60th": 538, "70th": 730, "80th": 1015, "90th": 1688, "95th": 2855, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00351-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108377, "hashes": {}}, "samples": 21352, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8098926, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108632, "hashes": {}}, "samples": 19656, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7868473, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108400, "hashes": {}}, "samples": 21766, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8175232, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 33154850, "hashes": {}}, "samples": 9565, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 4380395, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 57468679, "total_tokens_skipped": 0, "percentiles": {"0th": 59, "10th": 101, "20th": 146, "30th": 221, "40th": 295, "50th": 380, "60th": 540, "70th": 733, "80th": 1028, "90th": 1721, "95th": 2859, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00399-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105139, "hashes": {}}, "samples": 19600, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12880931, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67090883, "hashes": {}}, "samples": 18619, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12909399, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 13662397, "hashes": {}}, "samples": 4468, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3308351, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00399-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36284997, "total_tokens_skipped": 45, "percentiles": {"0th": 60, "10th": 101, "20th": 151, "30th": 231, "40th": 303, "50th": 391, "60th": 549, "70th": 760, "80th": 1088, "90th": 1814, "95th": 3129, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00399-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00421-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4caae8c3b05dc56b551227193fe23b04efe5a5f09c619330943e274da59a166b
3
+ size 48388663
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00425-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f4101a43de869cc08b2d5ad61b6bff6a0505b7129be9ee99d2ad327dfda5ce6
3
+ size 67105876
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0184-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c206e28bba7ec230f613c7cf768b1e7616d1d53ec97d8ee2e8f30408c0d75ee
3
+ size 46555644
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec1921edbbc9dda485861cbb277b22df9aedc54edfc27dcb70300783b19475df
3
+ size 67098323
train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00012-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75d525080121ffd8fa9423ff924ff7a3568376ae835210fe0200a539659de59b
3
+ size 5823413
train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84a27b6fa87aa7e16f5fc5fe4b20b4bae2ff69d582b75f778115afa0ca945a05
3
+ size 67097165
train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dfe3606275735b88346c84952f2539861892e0b8b244fe3ea05ad38fe27b749
3
+ size 64929924