Datasets:

ArXiv:
License:
orionweller commited on
Commit
a732ca3
·
verified ·
1 Parent(s): a23a310

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00160-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  2. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00211-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  3. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00221-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  4. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00234-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  5. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00280-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  6. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00364-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  7. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00369-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  8. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00369-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  9. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00369-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  10. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  11. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00471-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  12. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00576-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  13. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00664-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  14. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00664-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  15. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00664-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  16. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00674-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  17. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00730-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  18. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00869-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  19. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00869-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  20. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00869-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  21. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00874-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  22. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00974-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  23. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  24. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  25. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  26. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  27. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  28. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00173-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  29. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00205-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  30. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00240-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  31. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00285-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  32. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00293-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  33. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00293-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  34. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00293-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  35. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00310-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  36. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00310-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  37. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00320-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  38. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00375-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  39. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  40. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  41. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  42. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  43. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  44. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  45. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  46. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  47. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  48. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  49. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  50. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00160-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb1447543a034b8117a8b66116dc7443aa12b001a132c3a7d6270343ba77d81
3
+ size 3446817
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00211-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2641f14e9c18f177985764e4aeb28a227537342d83d55bb088c560e71a6c3196
3
+ size 25235388
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00221-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc52a50e35cc88215176520af90d63793e17c8b56fde9639f688497adc0c6a35
3
+ size 24580003
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00234-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3e4921041a38f022067d602d9dbad7ae3365ca564908312368d52e7909365fc
3
+ size 67107068
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00280-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59890e59630fc2ecff6dd3de0ee739e3698ffcdf71e304b9bd3c8fc1dac4755c
3
+ size 67105528
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00364-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a96e836f7d9b8110c858fae75ff9942fc4534fb4f0e3b2d9edbb4cd7b24ac64
3
+ size 26268243
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00369-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106973, "hashes": {}}, "samples": 17942, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 27002075, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13969931, "hashes": {}}, "samples": 3917, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6168015, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00369-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 19920372, "total_tokens_skipped": 25, "percentiles": {"0th": 57, "10th": 199, "20th": 267, "30th": 327, "40th": 389, "50th": 484, "60th": 638, "70th": 836, "80th": 1184, "90th": 1946, "95th": 3011, "99th": 7888, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00369-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae5a989b2fa10ca7a42a9f4fb7bd3007677083b4f605163714a50706b0b4da78
3
+ size 67108648
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00471-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f5e4c9ac0c2b181aa12b98e91649508dfb7ad18864e30872f4108d84489da8d
3
+ size 61100794
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00576-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c6333ccb1dfbbbfe0f8fce622ead62e02b0be7cb5b1e67aaffc053f07d09608
3
+ size 67105731
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00664-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107310, "hashes": {}}, "samples": 18301, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 20739849, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 41178600, "hashes": {}}, "samples": 11378, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12965031, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00664-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 26597919, "total_tokens_skipped": 13, "percentiles": {"0th": 62, "10th": 195, "20th": 262, "30th": 324, "40th": 386, "50th": 481, "60th": 638, "70th": 852, "80th": 1192, "90th": 1930, "95th": 2930, "99th": 7798, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00664-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00674-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55b8e5bff4e0b896c77b34130e9792349d70cd513178be87ff627b038e05a015
3
+ size 58070559
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00730-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64f3cefa2a88ec9ccf8f9222f9dbb05d64ae9eca75dafee7d503d67fe6156d0d
3
+ size 22361330
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00869-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108825, "hashes": {}}, "samples": 20423, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11656577, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67104495, "hashes": {}}, "samples": 18757, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13701358, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 16695826, "hashes": {}}, "samples": 4946, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3206581, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00869-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 37023248, "total_tokens_skipped": 43, "percentiles": {"0th": 62, "10th": 193, "20th": 258, "30th": 319, "40th": 380, "50th": 462, "60th": 609, "70th": 795, "80th": 1096, "90th": 1772, "95th": 2677, "99th": 7032, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00869-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00874-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:231e2f755f59e64f6a7ea0d70128841b07cb44bbd61d5ca1f64f5fd42aec13ef
3
+ size 17059690
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00974-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c386ff97dd6e6ea1f924e2f29e4c96d3290040f3f4009e0b617aed645ef7bc55
3
+ size 37557963
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67100745, "hashes": {}}, "samples": 16979, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7139853, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67087894, "hashes": {}}, "samples": 18447, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7284977, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107638, "hashes": {}}, "samples": 18587, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6772692, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108647, "hashes": {}}, "samples": 20237, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7464759, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 1697594, "hashes": {}}, "samples": 554, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 174876, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 66334525, "total_tokens_skipped": 145, "percentiles": {"0th": 61, "10th": 106, "20th": 158, "30th": 237, "40th": 311, "50th": 400, "60th": 550, "70th": 763, "80th": 1098, "90th": 1940, "95th": 3431, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98972385722db9f136dfdd7e5eceb9079e4fd730e6e9475780683e5d5caa81d3
3
+ size 67106345
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71ee0935319f033838779c254da44c73a9083d75f1414cffa331741a58a62c34
3
+ size 67108707
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00173-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4eecbd44f0da476c193d983d3c67c4c8313c9327002dffe8e7a2a9a0c011e44
3
+ size 67107204
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00205-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2065e00591fb8fc7234a79a3e3696c4065c0f851cc8bc8a14ff58a405b90f3af
3
+ size 67108563
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00240-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7090f715701255fbd33dca0efda23fa84b707103b219a520d841e7f3d8769c28
3
+ size 24401126
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00285-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f04ade5605644710d11062fbc22fae178c2cc184795e9f15ac43de99de4dc3b6
3
+ size 67102165
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00293-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67093111, "hashes": {}}, "samples": 16555, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12343865, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102179, "hashes": {}}, "samples": 17068, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12990389, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 51536468, "hashes": {}}, "samples": 12207, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 10583593, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00293-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 45702955, "total_tokens_skipped": 123, "percentiles": {"0th": 59, "10th": 104, "20th": 164, "30th": 254, "40th": 335, "50th": 468, "60th": 660, "70th": 915, "80th": 1320, "90th": 2266, "95th": 3890, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00293-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00310-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b83e623ba75601673cf6df6c6de060ee2b56ffd46b32a2f71c3491f1c2283f3c
3
+ size 67107328
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00310-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e89e2018c7981663d4b50137ee68ec692c7e86bfa94924b4166c20bb416da422
3
+ size 67108276
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00320-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f488f1bebabe8f9647588f2f666e1d6f3b9ecd6e7e068793c49c5c87ee0b8e5
3
+ size 33983027
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00375-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84b1d5480d18258ec9750e6d60ad20648c47dba2d0155afe9e7b90c0febd0067
3
+ size 51485489
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc9cf39edd1ac4b82de240d4397a9d7fdd55ff52e8d69308885c990f515ca02d
3
+ size 67105405
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a439db0e63820f155569d694b40f5dc4dfea65e33dc10aaf7d22cdff34ee9ac
3
+ size 67103730
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cee6e65dc9976dfa3662e0bfc20ca5c52ae8eb942dc79410bfdcae8701e964ac
3
+ size 67100780
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67917ee0fa85f67b9c1bd7a92d4a01d29b3b8fb433af3614e27c22825c0da24b
3
+ size 67103947
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0716ff5b65a283bed584821bc1327efa7fbf6ed9607dde9db32d077a0b64cfc6
3
+ size 67108312
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0074c48af19c893ee153e0bedaa22a2e28a7046c497bd7e090a6f436215e608c
3
+ size 67108498
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a6252693abf1fab884b21b9873eb46d284ec6b6f60d469fc1d681f5bf701a5e
3
+ size 67105149
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67096505, "hashes": {}}, "samples": 13755, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11401652, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106270, "hashes": {}}, "samples": 13806, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11924342, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108620, "hashes": {}}, "samples": 14184, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 12214626, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 28415857, "hashes": {}}, "samples": 6511, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 4809461, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 56662362, "total_tokens_skipped": 152, "percentiles": {"0th": 87, "10th": 153, "20th": 222, "30th": 322, "40th": 413, "50th": 512, "60th": 710, "70th": 1003, "80th": 1556, "90th": 2891, "95th": 5177, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103980, "hashes": {}}, "samples": 11594, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16925082, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108015, "hashes": {}}, "samples": 11616, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16497051, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 44766166, "hashes": {}}, "samples": 7818, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11513827, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 44249495, "total_tokens_skipped": 52, "percentiles": {"0th": 87, "10th": 161, "20th": 253, "30th": 365, "40th": 459, "50th": 604, "60th": 850, "70th": 1233, "80th": 1969, "90th": 3942, "95th": 6796, "99th": 8191, "100th": 8191}}