Datasets:

ArXiv:
License:
orionweller commited on
Commit
66865a3
·
verified ·
1 Parent(s): e133c1d

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00067-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  2. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00133-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  3. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00133-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  4. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00133-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  5. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  6. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  7. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00189-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  8. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00189-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  9. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00196-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  10. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00213-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  11. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  12. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  13. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  14. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  15. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  16. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00324-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  17. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  18. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00361-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  19. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00375-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  20. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00375-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  21. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00378-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  22. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00380-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  23. train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0022-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  24. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  25. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  26. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  27. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  28. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  29. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  30. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  31. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00004-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  32. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00004-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  33. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00004-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  34. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  35. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  36. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  37. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  38. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  39. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  40. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  41. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  42. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  43. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  44. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  45. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  46. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  47. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  48. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  49. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  50. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00067-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6c7e8011a0e356ff36c2d522134b9fbdeb06682d2d8563151f49f918ae38262
3
+ size 28552646
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00133-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108779, "hashes": {}}, "samples": 18394, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7899846, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102776, "hashes": {}}, "samples": 18384, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7980975, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107726, "hashes": {}}, "samples": 17554, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7352067, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108241, "hashes": {}}, "samples": 18309, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 8153231, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 8950358, "hashes": {}}, "samples": 2543, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 940391, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00133-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 68147575, "total_tokens_skipped": 176, "percentiles": {"0th": 61, "10th": 103, "20th": 151, "30th": 229, "40th": 301, "50th": 383, "60th": 540, "70th": 764, "80th": 1105, "90th": 2032, "95th": 3775, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00133-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9d3133dbc7f6c9b079db1593da54dd8d7f8adb57a7eac89043bea7e1cd1a265
3
+ size 67108051
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b1ba25fef8e76af9871d7ab2be7d8f57db1321583c2f66f074331e7ab5c64cf
3
+ size 67086171
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00189-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67081753, "hashes": {}}, "samples": 17568, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11820037, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105918, "hashes": {}}, "samples": 16456, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12290957, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 66032397, "hashes": {}}, "samples": 15631, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11941562, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00189-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00196-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79c8af149cde812170e6f1d4f7ac774e2e3497266ea56ccdb4233aaa6abea1c8
3
+ size 58822037
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00213-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2899f7ea010620a93bc2fc44326aa142fd8e21d5f14635808a8c834008e3fbc9
3
+ size 67108017
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00215-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:441b0b3b31ed69670f92d12ab03d7c777bd77fce22110cbd22d9bebd16ec6257
3
+ size 43441222
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103818, "hashes": {}}, "samples": 20198, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9889635, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108600, "hashes": {}}, "samples": 21080, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10436794, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67105603, "hashes": {}}, "samples": 21936, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6887023, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 6504775, "hashes": {}}, "samples": 2394, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 666774, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 50911587, "total_tokens_skipped": 19, "percentiles": {"0th": 61, "10th": 101, "20th": 147, "30th": 217, "40th": 288, "50th": 366, "60th": 503, "70th": 691, "80th": 978, "90th": 1627, "95th": 2749, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00247-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00324-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:001dc6d61ec575e74f1f62c63c5ab48a3c9d1196496ab720c597049402af6c26
3
+ size 26516060
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00360-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0f0c7329389d5bb6811186aa51a67bfe9c0bbf5d858772970d3180b1ac08825
3
+ size 67104685
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00361-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80e6a00b5c5738fc39473c259a6108cb19b30251b439963610819e7fff3b6ae1
3
+ size 67107891
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00375-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 45500714, "total_tokens_skipped": 115, "percentiles": {"0th": 62, "10th": 102, "20th": 149, "30th": 222, "40th": 296, "50th": 377, "60th": 531, "70th": 722, "80th": 1000, "90th": 1669, "95th": 2729, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00375-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00378-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864cd68e1038dab99a2004ed5d1025569611d043eda98d549935c544fbebea8f
3
+ size 67108663
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00380-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:646352061d8e8c8bf558ed9d947cbaeae92718c7d932d377427cbcf01bd9da86
3
+ size 67102947
train/fineweb2-sampled-ext-v2/guj_Gujr_train-sampled/batch_0022-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107086, "hashes": {}}, "samples": 11099, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 4701937, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 63130567, "hashes": {}}, "samples": 9180, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5387620, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108156, "hashes": {}}, "samples": 16915, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5945324, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67095372, "hashes": {}}, "samples": 14392, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6529310, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67082315, "hashes": {}}, "samples": 12457, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7056221, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108797, "hashes": {}}, "samples": 14693, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 8008725, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67088319, "hashes": {}}, "samples": 13464, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 6338781, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 57923388, "hashes": {}}, "samples": 13868, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 5131629, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 96983905, "total_tokens_skipped": 438, "percentiles": {"0th": 90, "10th": 152, "20th": 216, "30th": 308, "40th": 389, "50th": 474, "60th": 628, "70th": 892, "80th": 1394, "90th": 2740, "95th": 5328, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101821, "hashes": {}}, "samples": 13167, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8310544, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102437, "hashes": {}}, "samples": 13177, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6824787, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67106942, "hashes": {}}, "samples": 18596, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6498042, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 8013027, "hashes": {}}, "samples": 1753, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 834925, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb4832087dc593a7de8de70adcb01b8b3cb0f47dfdf9a21c0d3e7cd1b3ae98f
3
+ size 8013027
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 51586572, "total_tokens_skipped": 90, "percentiles": {"0th": 85, "10th": 155, "20th": 221, "30th": 310, "40th": 391, "50th": 473, "60th": 617, "70th": 885, "80th": 1394, "90th": 2668, "95th": 4796, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00004-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67091120, "hashes": {}}, "samples": 16554, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6184845, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106875, "hashes": {}}, "samples": 13974, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6952511, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107921, "hashes": {}}, "samples": 13689, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7785093, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67105600, "hashes": {}}, "samples": 12691, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7810966, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67105776, "hashes": {}}, "samples": 15937, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5872917, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 50467893, "hashes": {}}, "samples": 11866, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 4537572, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00004-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 95145763, "total_tokens_skipped": 185, "percentiles": {"0th": 76, "10th": 154, "20th": 216, "30th": 308, "40th": 389, "50th": 475, "60th": 627, "70th": 908, "80th": 1401, "90th": 2697, "95th": 5094, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00004-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa3cc95dcf61a30208b3dba2550dac1408782de3e8bb2cba518945a5ca9db275
3
+ size 38313854
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67096049, "hashes": {}}, "samples": 15155, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8369258, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108569, "hashes": {}}, "samples": 13239, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9965660, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 20279238, "hashes": {}}, "samples": 5298, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 2218316, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38083780, "total_tokens_skipped": 58, "percentiles": {"0th": 92, "10th": 152, "20th": 214, "30th": 308, "40th": 383, "50th": 474, "60th": 633, "70th": 899, "80th": 1447, "90th": 2794, "95th": 5052, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107836, "hashes": {}}, "samples": 17289, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8070667, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107956, "hashes": {}}, "samples": 13092, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9331512, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67080804, "hashes": {}}, "samples": 15777, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7863561, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67104881, "hashes": {}}, "samples": 12664, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 9328208, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 35903905, "hashes": {}}, "samples": 8735, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 4482955, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 74999403, "total_tokens_skipped": 145, "percentiles": {"0th": 85, "10th": 151, "20th": 212, "30th": 300, "40th": 383, "50th": 469, "60th": 618, "70th": 895, "80th": 1402, "90th": 2730, "95th": 4988, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00007-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94205df00b53c702ae8911d606bffdcac17b8fc8ca99e4767b2fb1d7a130c368
3
+ size 27500768
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67081492, "hashes": {}}, "samples": 16203, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8937490, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 66645727, "hashes": {}}, "samples": 12432, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9872274, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 32975240, "total_tokens_skipped": 0, "percentiles": {"0th": 88, "10th": 153, "20th": 220, "30th": 314, "40th": 402, "50th": 493, "60th": 678, "70th": 970, "80th": 1521, "90th": 2798, "95th": 4970, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ccba2b69247d5c7026b957dcc1130cb1751d9fe64b928223d186c4a679ec895
3
+ size 31331780
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103244, "hashes": {}}, "samples": 11201, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16334246, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103755, "hashes": {}}, "samples": 11237, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16200862, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 53506024, "hashes": {}}, "samples": 9366, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 13016788, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 46420868, "total_tokens_skipped": 17, "percentiles": {"0th": 88, "10th": 161, "20th": 248, "30th": 367, "40th": 461, "50th": 611, "60th": 878, "70th": 1267, "80th": 2022, "90th": 4093, "95th": 7217, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105510, "hashes": {}}, "samples": 14542, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7234398, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67097677, "hashes": {}}, "samples": 13554, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8276413, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67098887, "hashes": {}}, "samples": 13993, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7308917, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108674, "hashes": {}}, "samples": 16864, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6272186, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67105241, "hashes": {}}, "samples": 14292, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 7297504, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 29543047, "hashes": {}}, "samples": 5644, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 3735823, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d90d911cb01c7f25cea04a14478691d2a028e1815851a8103eeed18e211db75d
3
+ size 29543047