Datasets:

ArXiv:
License:
orionweller commited on
Commit
d075b35
·
verified ·
1 Parent(s): 30e3060

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  2. train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  3. train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0069-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  4. train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  5. train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0113-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  6. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  7. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  8. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  9. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  10. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  11. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  12. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  13. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  14. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  15. train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
  16. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0097-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  17. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  18. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0190-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  19. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0493-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  20. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0493-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  21. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0493-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  22. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0502-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  23. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0502-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  24. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0502-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  25. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0526-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  26. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0526-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  27. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0526-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  28. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0603-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  29. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0633-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  30. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0883-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  31. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0883-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  32. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0883-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  33. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1198-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  34. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1198-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  35. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1198-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  36. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1217-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  37. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1251-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  38. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1251-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  39. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1352-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  40. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1410-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  41. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1410-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  42. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1410-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  43. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1486-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  44. train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1527-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  45. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00100-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  46. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00100-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  47. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00100-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  48. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00126-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  49. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  50. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00179-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108515, "hashes": {}}, "samples": 20838, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5972295, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105068, "hashes": {}}, "samples": 23697, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6018446, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67106080, "hashes": {}}, "samples": 24847, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6118951, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67105512, "hashes": {}}, "samples": 24306, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6120966, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 510668, "hashes": {}}, "samples": 272, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 59011, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 65737396, "total_tokens_skipped": 155, "percentiles": {"0th": 66, "10th": 134, "20th": 192, "30th": 255, "40th": 331, "50th": 424, "60th": 542, "70th": 699, "80th": 931, "90th": 1397, "95th": 2035, "99th": 5544, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0069-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dedf2ab457098af1399aa5f7166f73d83babc209541f7c74b064081c761a4c8f
3
+ size 67108561
train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8c9549b87037253402c7395ed5fc4d5191173e7b25a433655a60684622cde94
3
+ size 67105166
train/fineweb2-sampled-ext-v2/cat_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/002_00000-batch_0113-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f04b27a83cfdc07d8ee474223aec3138a63264e1409bd468e511d3f150cd43a
3
+ size 67107519
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd2b9dc1f609515ef525f5702856397889d1d524c084e1cbc2368a2176d51346
3
+ size 67102869
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00014-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4887906b4492dc05861d11ee17925db4fa15c46c93b164affad113e6ab481866
3
+ size 13395785
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0ad995117d57adececc770c95074ccd2fd644734fb3d05e0bfd1f4f4df78daa
3
+ size 67108155
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d720dae31f4242da98c4b0cbe3b18fb57edd1f8b1c5b76fa7e0a4c3cc6d6a921
3
+ size 67106747
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00027-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cc06a94ba9b152f3b9494af017d9b991d999044ea23d896159bc789d8c6b411
3
+ size 67095891
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 46515582, "total_tokens_skipped": 133, "percentiles": {"0th": 98, "10th": 177, "20th": 266, "30th": 390, "40th": 512, "50th": 675, "60th": 956, "70th": 1354, "80th": 2152, "90th": 4508, "95th": 8190, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102206, "hashes": {}}, "samples": 13756, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6564925, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67093904, "hashes": {}}, "samples": 12764, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7568760, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107623, "hashes": {}}, "samples": 12923, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5992277, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108591, "hashes": {}}, "samples": 15339, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5813009, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67105514, "hashes": {}}, "samples": 11401, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 7187655, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 57890236, "hashes": {}}, "samples": 9599, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 6413183, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00034-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91633286c663f443831c0c84506b52478acad8bf474718ee398d809ebc37b9a0
3
+ size 17943294
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0097-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ce76300c6bdf71fd0c441253a00de77eb7c2b1186efef5294451d5b1ba2c388
3
+ size 67104894
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0189-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d27bd8fd1434455b47aa7ade98d2a9aa7405a4d39c0ffc5b26452524905d949
3
+ size 67106943
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0190-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0493-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 66125677, "hashes": {}}, "samples": 20154, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15373150, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0493-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 16209910, "total_tokens_skipped": 0, "percentiles": {"0th": 91, "10th": 194, "20th": 255, "30th": 319, "40th": 400, "50th": 510, "60th": 641, "70th": 817, "80th": 1056, "90th": 1589, "95th": 2244, "99th": 6417, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0493-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0502-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 65618497, "hashes": {}}, "samples": 20147, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15149358, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0502-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 16083260, "total_tokens_skipped": 0, "percentiles": {"0th": 84, "10th": 190, "20th": 251, "30th": 321, "40th": 398, "50th": 504, "60th": 627, "70th": 800, "80th": 1048, "90th": 1604, "95th": 2305, "99th": 6016, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0502-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0526-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106966, "hashes": {}}, "samples": 15148, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16267701, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18272138, "hashes": {}}, "samples": 5177, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4240753, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0526-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21020292, "total_tokens_skipped": 0, "percentiles": {"0th": 101, "10th": 210, "20th": 285, "30th": 365, "40th": 471, "50th": 599, "60th": 770, "70th": 999, "80th": 1357, "90th": 2129, "95th": 3419, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0526-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0603-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a755281678bc246a58e978b251685a31ecfc8316d2ae595eba608be7ce34a71b
3
+ size 67107482
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0633-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed73d57428f16cea94329ff295f7791090545561e53b42711ca0e5d1b42da856
3
+ size 24813154
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0883-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106051, "hashes": {}}, "samples": 19725, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14694034, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 2278048, "hashes": {}}, "samples": 547, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 543843, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0883-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 17022582, "total_tokens_skipped": 0, "percentiles": {"0th": 88, "10th": 195, "20th": 255, "30th": 318, "40th": 398, "50th": 500, "60th": 631, "70th": 804, "80th": 1052, "90th": 1615, "95th": 2405, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_0883-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1198-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108713, "hashes": {}}, "samples": 16031, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15367939, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16043122, "hashes": {}}, "samples": 4134, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3545946, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1198-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 20466019, "total_tokens_skipped": 0, "percentiles": {"0th": 84, "10th": 208, "20th": 285, "30th": 373, "40th": 491, "50th": 646, "60th": 822, "70th": 1049, "80th": 1410, "90th": 2213, "95th": 2980, "99th": 7101, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1198-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1217-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24d74288176a17670a10b355d92261e7ed6228df1dc7309c6e6de90be91233b8
3
+ size 67108139
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1251-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106038, "hashes": {}}, "samples": 17835, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15258640, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8574600, "hashes": {}}, "samples": 2342, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 1936892, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1251-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1352-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06539704cc29676b7cdb02a6ab08d4eceabd29e56e4198ab9a9604576f54c130
3
+ size 67107655
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1410-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105750, "hashes": {}}, "samples": 17466, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14837384, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10484160, "hashes": {}}, "samples": 2724, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2277370, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1410-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 19075191, "total_tokens_skipped": 36, "percentiles": {"0th": 82, "10th": 208, "20th": 278, "30th": 363, "40th": 465, "50th": 604, "60th": 761, "70th": 956, "80th": 1266, "90th": 1894, "95th": 2771, "99th": 7737, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1410-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1486-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:829168c5266277e3d43deb241c21657caefc0d4ae78f200b2d25dfa2565b401e
3
+ size 67105561
train/fineweb2-sampled-ext-v2/fin_Latn_train-sampled/batch_1527-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af56e00b6cc35128da9a0a4f3a9fac802e2750cf51b2694b527b53f5c24a2ec2
3
+ size 33885540
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00100-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102186, "hashes": {}}, "samples": 18009, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14568365, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103137, "hashes": {}}, "samples": 18128, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15779504, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 3477237, "hashes": {}}, "samples": 855, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 797223, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00100-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 33831303, "total_tokens_skipped": 0, "percentiles": {"0th": 59, "10th": 106, "20th": 172, "30th": 263, "40th": 349, "50th": 494, "60th": 679, "70th": 921, "80th": 1286, "90th": 2035, "95th": 3142, "99th": 7859, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00100-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00126-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c536c453950825336cd1f8e58e0b926ccdc667301add999903d59f661d5160a1
3
+ size 32662532
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:164fae7077d2e95461a149aa8dc319a9419b91038d217ece2d8c7b78a1faf139
3
+ size 15911135
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00179-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09fb6681be9ad7c96bd75d2d7be063a11b2ab3a15ef628e4a44276c523be0933
3
+ size 67093721