Datasets:

ArXiv:
License:
orionweller commited on
Commit
ab5836a
·
verified ·
1 Parent(s): 8673e3e

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  2. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
  3. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  4. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  5. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  6. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  7. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  8. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  9. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0028-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  10. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0028-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  11. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0028-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  12. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  13. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  14. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  15. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  16. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  17. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  18. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  19. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  20. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  21. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  22. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  23. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0061-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  24. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0061-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  25. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0061-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  26. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0064-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  27. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0064-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  28. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0064-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  29. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  30. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  31. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  32. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  33. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0076-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  34. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0076-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  35. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0076-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  36. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0077-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  37. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0077-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  38. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0077-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  39. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  40. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  41. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  42. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0099-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  43. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0099-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  44. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0099-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  45. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  46. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  47. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  48. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0122-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  49. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0142-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  50. train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0142-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101493, "hashes": {}}, "samples": 17546, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6663994, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105893, "hashes": {}}, "samples": 11222, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5933510, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107186, "hashes": {}}, "samples": 16472, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5064371, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67100655, "hashes": {}}, "samples": 18360, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5392619, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67104654, "hashes": {}}, "samples": 17634, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5610895, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 594811, "hashes": {}}, "samples": 191, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 58784, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7d9f9c58653ec975e9ae98a58f084a84bad31fc7ed9bc6a881ec58346753251
3
+ size 594811
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 82729722, "total_tokens_skipped": 0, "percentiles": {"0th": 83, "10th": 191, "20th": 252, "30th": 313, "40th": 388, "50th": 496, "60th": 645, "70th": 902, "80th": 1313, "90th": 2297, "95th": 3784, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67091513, "hashes": {}}, "samples": 16664, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6578243, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106492, "hashes": {}}, "samples": 12619, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6483459, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67105574, "hashes": {}}, "samples": 17207, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5075377, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108324, "hashes": {}}, "samples": 18637, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5429397, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 60894928, "hashes": {}}, "samples": 14782, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5046774, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 81052097, "total_tokens_skipped": 0, "percentiles": {"0th": 92, "10th": 189, "20th": 250, "30th": 312, "40th": 386, "50th": 495, "60th": 643, "70th": 896, "80th": 1329, "90th": 2277, "95th": 3784, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19af43696f845d8b1acf367484dc668b6a695eeb01bd5e6f6f33ae6a0e9a8b3c
3
+ size 43159078
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0028-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108600, "hashes": {}}, "samples": 17850, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6470675, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105535, "hashes": {}}, "samples": 17610, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6854262, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107004, "hashes": {}}, "samples": 12216, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5991235, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108222, "hashes": {}}, "samples": 20603, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5438502, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 24757490, "hashes": {}}, "samples": 6760, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 2124770, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0028-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 72099674, "total_tokens_skipped": 0, "percentiles": {"0th": 92, "10th": 192, "20th": 252, "30th": 315, "40th": 386, "50th": 483, "60th": 625, "70th": 848, "80th": 1241, "90th": 2114, "95th": 3455, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0028-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4149ef87572133a4898cfa7957d488b6eb236c73b1919036886227b35acbe941
3
+ size 67108079
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e0c6f301bfd368c525f7f08d1a6765fa795943822089b6937868644fbffd75c
3
+ size 67082835
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8fecc90b7801d15e4268268e966d2dff1cd1d3dabd6fd104f0fe17313a7e8a5
3
+ size 67104561
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0038-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4012e22d2a86474f1f17874293549d731adceb4ab7657c20cb11a255c7c4a1d8
3
+ size 58746815
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f79e93fb331006e03592d25e64f24f6b573c61375f65b01a3e1e3883e27adeff
3
+ size 67076328
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4199729835a74d166624e6192a2a5bd4399d9cad610422b403bfb74bfb81497c
3
+ size 67091809
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:926f85709f4c8bff41bdf81f4214a7a7e517a583759bf65f48fc620287442c9f
3
+ size 67107614
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:542e0d21bfcbf4e3c05394bdb4ae6f9d7a4ac0f5d0e2c5aa4f039fff35dc2902
3
+ size 67106814
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67082693, "hashes": {}}, "samples": 19600, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5995914, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67094297, "hashes": {}}, "samples": 18018, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6660066, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67104895, "hashes": {}}, "samples": 17411, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7165515, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108790, "hashes": {}}, "samples": 16388, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5724818, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 1180773, "hashes": {}}, "samples": 228, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 101244, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 66249870, "total_tokens_skipped": 4, "percentiles": {"0th": 90, "10th": 196, "20th": 255, "30th": 315, "40th": 387, "50th": 476, "60th": 603, "70th": 809, "80th": 1178, "90th": 2031, "95th": 3316, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0061-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67100856, "hashes": {}}, "samples": 20511, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5614588, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108620, "hashes": {}}, "samples": 18991, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6606633, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108826, "hashes": {}}, "samples": 17838, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6936262, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 63298095, "hashes": {}}, "samples": 14088, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6374553, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0061-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 65014851, "total_tokens_skipped": 0, "percentiles": {"0th": 92, "10th": 190, "20th": 247, "30th": 303, "40th": 369, "50th": 456, "60th": 579, "70th": 783, "80th": 1151, "90th": 1946, "95th": 3278, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0061-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0064-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67100290, "hashes": {}}, "samples": 19000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5801240, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67091313, "hashes": {}}, "samples": 18688, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6795396, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108852, "hashes": {}}, "samples": 18195, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7471670, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 59664546, "hashes": {}}, "samples": 15369, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5255952, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0064-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 64104789, "total_tokens_skipped": 150, "percentiles": {"0th": 77, "10th": 193, "20th": 248, "30th": 307, "40th": 372, "50th": 456, "60th": 585, "70th": 796, "80th": 1163, "90th": 1946, "95th": 3095, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0064-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86a72e0321bf0a6505668381fa1c686cd6d4d7e982ec50b695875a4667c8bb0c
3
+ size 67082171
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89e4fb3b103f9b29744c6c36cdca53971819bea995a8fdf0b37532f16d3aef50
3
+ size 67107353
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02cda30f85e4b58f82512a476e5573b76d04c505ab2da88a2258d73e6a327bdc
3
+ size 67105261
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dde5140ac3eac85a07b0c2b97ddec06c4258e8b568172f8cbd142e1930b08bb
3
+ size 40477445
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0076-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108653, "hashes": {}}, "samples": 19763, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6132807, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108526, "hashes": {}}, "samples": 19184, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6547722, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108547, "hashes": {}}, "samples": 17951, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7245997, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 49117022, "hashes": {}}, "samples": 12247, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5229413, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0076-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 61507847, "total_tokens_skipped": 81, "percentiles": {"0th": 83, "10th": 190, "20th": 248, "30th": 307, "40th": 376, "50th": 466, "60th": 591, "70th": 788, "80th": 1143, "90th": 1928, "95th": 3083, "99th": 8115, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0076-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0077-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107689, "hashes": {}}, "samples": 20717, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5885356, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67098918, "hashes": {}}, "samples": 19542, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7101338, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67105441, "hashes": {}}, "samples": 17815, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7777767, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 44685340, "hashes": {}}, "samples": 13003, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 3731148, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0077-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 60365958, "total_tokens_skipped": 85, "percentiles": {"0th": 85, "10th": 187, "20th": 240, "30th": 297, "40th": 360, "50th": 442, "60th": 565, "70th": 753, "80th": 1105, "90th": 1830, "95th": 2903, "99th": 7332, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0077-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108362, "hashes": {}}, "samples": 19063, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6903264, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102792, "hashes": {}}, "samples": 17388, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7313341, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67103405, "hashes": {}}, "samples": 18793, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6312549, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 44102473, "hashes": {}}, "samples": 12834, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 4308922, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 60268365, "total_tokens_skipped": 34, "percentiles": {"0th": 88, "10th": 194, "20th": 250, "30th": 310, "40th": 375, "50th": 459, "60th": 577, "70th": 768, "80th": 1111, "90th": 1891, "95th": 3048, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0086-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0099-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c9e9db1255f281f35cbe2686892847cbf13bca077d2af3ca5c0f056db91bc9e
3
+ size 67106573
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0099-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95e94aafce4a243194db3e77e7a0511cfab6aed948a12b26a7b205921f8f79f6
3
+ size 67107538
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0099-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cef4e2f26e91971b58609d2b67c5dc54da7fe27b1099331ab6db5c4729217308
3
+ size 67108474
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0107-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72cdee4dc28d1c143183762a3c9f7069ed06b13b58236291d48b6c8ca506781f
3
+ size 26838143
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b19daa5ef842360a213b331f292b99b228df70d8cd80c00cdbdfc61e52dcd65
3
+ size 67107939
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f95f83a682a2d9d16fd4a4b58abf11a33585b8af47338aeb470840ca4be7489
3
+ size 67107427
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0122-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d284a81c658069928a4b06d06cfc7503c00b6e2021c2f237f58d63f1b06340c0
3
+ size 4169208
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0142-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108439, "hashes": {}}, "samples": 19876, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7557141, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67096835, "hashes": {}}, "samples": 20518, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8039763, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 66732027, "hashes": {}}, "samples": 20162, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7438949, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/als_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0142-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 49268680, "total_tokens_skipped": 0, "percentiles": {"0th": 88, "10th": 188, "20th": 242, "30th": 297, "40th": 363, "50th": 444, "60th": 550, "70th": 724, "80th": 1027, "90th": 1710, "95th": 2692, "99th": 6937, "100th": 8191}}