Datasets:

ArXiv:
License:
orionweller commited on
Commit
c5f30a4
·
verified ·
1 Parent(s): 83e71d6

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0135-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  2. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/004_00000-batch_0022-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  3. train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/005_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  4. train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  5. train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  6. train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0014-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  7. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  8. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  9. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  10. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  11. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00542-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  12. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00542-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  13. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00542-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  14. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00687-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  15. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00862-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  16. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00862-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  17. train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00862-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  18. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00006.mds +3 -0
  19. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00124-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  20. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  21. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  22. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00200-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  23. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00240-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  24. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00258-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  25. train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00318-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  26. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  27. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  28. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  29. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  30. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  31. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  32. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  33. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  34. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  36. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  37. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00038-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  38. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00038-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  39. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00038-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  40. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00039-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  41. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00040-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  42. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00044-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  43. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00044-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  44. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  45. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  46. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  47. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  48. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  49. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00052-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  50. train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00052-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/003_00000-batch_0135-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6071eb49738b1a0cb2a422ebb0de39d7c755d25d37ab4155ad3b6aca09b0eba1
3
+ size 67095454
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/004_00000-batch_0022-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fa2e25b02c7dfb39f78885ce6c5bdfc2102b46a8cd6ab6231aa1bf61028eccd
3
+ size 67107759
train/fineweb2-sampled-ext-v2/bos_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/005_00000-batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e02639c996c4d157fb85f64f559b041c97880165d0b70474302d88c010a6132e
3
+ size 67096329
train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106629, "hashes": {}}, "samples": 10294, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6711489, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67097355, "hashes": {}}, "samples": 11137, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6680947, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67092373, "hashes": {}}, "samples": 10347, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6661203, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67100799, "hashes": {}}, "samples": 11674, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6774780, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67092801, "hashes": {}}, "samples": 10936, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 6754428, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 10056506, "hashes": {}}, "samples": 1979, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 921472, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0012-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0014-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101506, "hashes": {}}, "samples": 11445, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7119045, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67084155, "hashes": {}}, "samples": 10426, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7292388, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67083701, "hashes": {}}, "samples": 10523, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7367721, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108307, "hashes": {}}, "samples": 11106, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7458198, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 44138421, "hashes": {}}, "samples": 7241, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5206765, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00036-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104036, "hashes": {}}, "samples": 18635, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7176764, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67091178, "hashes": {}}, "samples": 17343, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7541568, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67093774, "hashes": {}}, "samples": 18112, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7299673, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67105546, "hashes": {}}, "samples": 17512, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7389391, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 45221250, "hashes": {}}, "samples": 12876, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5478604, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 63167096, "hashes": {}}, "samples": 18168, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16386595, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 15501873, "total_tokens_skipped": 18, "percentiles": {"0th": 59, "10th": 199, "20th": 266, "30th": 323, "40th": 380, "50th": 468, "60th": 615, "70th": 794, "80th": 1131, "90th": 1814, "95th": 2783, "99th": 6776, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00542-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106575, "hashes": {}}, "samples": 17758, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15602519, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103315, "hashes": {}}, "samples": 18760, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 17001726, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 13827521, "hashes": {}}, "samples": 3461, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3024727, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00542-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 36371165, "total_tokens_skipped": 35, "percentiles": {"0th": 60, "10th": 206, "20th": 281, "30th": 345, "40th": 414, "50th": 524, "60th": 677, "70th": 878, "80th": 1217, "90th": 1957, "95th": 2817, "99th": 7485, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00542-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00687-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40cb33eb7a9a66d75b1710bc00ed6fd146c7efe17ecd5ba59454ad954bb8e1fc
3
+ size 67077853
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00862-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105575, "hashes": {}}, "samples": 18264, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 22369040, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 49741608, "hashes": {}}, "samples": 14884, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11760032, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00862-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 28682807, "total_tokens_skipped": 17, "percentiles": {"0th": 62, "10th": 196, "20th": 268, "30th": 331, "40th": 394, "50th": 501, "60th": 654, "70th": 840, "80th": 1168, "90th": 1907, "95th": 2794, "99th": 5771, "100th": 8191}}
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00862-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26f77d9f43e8884bf3c31402c682439d6bd1d026bae83aaf4618d322acade95e
3
+ size 34103631
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00124-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107542, "hashes": {}}, "samples": 18411, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7297358, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108370, "hashes": {}}, "samples": 17348, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7171157, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67087788, "hashes": {}}, "samples": 18445, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6989009, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67104980, "hashes": {}}, "samples": 18730, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7744154, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 35789730, "hashes": {}}, "samples": 9504, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 3579304, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 74737067, "total_tokens_skipped": 55, "percentiles": {"0th": 60, "10th": 105, "20th": 153, "30th": 235, "40th": 309, "50th": 392, "60th": 537, "70th": 745, "80th": 1089, "90th": 2000, "95th": 3853, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00200-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:395f826fe593e58edbae15d5bedc6f41967c720a6e88419d5c20d461db935b71
3
+ size 67085019
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00240-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b197e5431447df6ccbc3f35bb985480424c767077caa0cfacbae6ee8d637d971
3
+ size 67078421
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00258-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d154dee19a371daa37cef5e765f297d29028618252b1a6342209a177059d87b7
3
+ size 61148183
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00318-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6fc266f2fe399b8f8b663158cd527dff3e73ad43fd5ed967048e270cedc8d5a
3
+ size 45729871
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deb95c879a0576dbb132008b66caba9603e8638ca1bf94d30eb98b125c6e5778
3
+ size 14904547
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105269, "hashes": {}}, "samples": 13616, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9071848, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67078772, "hashes": {}}, "samples": 15025, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8453354, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107297, "hashes": {}}, "samples": 13352, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9095177, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67090392, "hashes": {}}, "samples": 15433, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 8352898, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 49173377, "hashes": {}}, "samples": 9103, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 6515150, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 78327927, "total_tokens_skipped": 35, "percentiles": {"0th": 89, "10th": 156, "20th": 226, "30th": 320, "40th": 401, "50th": 491, "60th": 659, "70th": 937, "80th": 1462, "90th": 2894, "95th": 5564, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00009-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:758f5489315da270696b85af01b4a0861caae7647123c0bc32482489902d07c1
3
+ size 29876894
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105972, "hashes": {}}, "samples": 12505, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16167844, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67098323, "hashes": {}}, "samples": 11294, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15990408, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 48663938, "hashes": {}}, "samples": 8863, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11710624, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 45195980, "total_tokens_skipped": 149, "percentiles": {"0th": 84, "10th": 160, "20th": 250, "30th": 358, "40th": 451, "50th": 588, "60th": 829, "70th": 1210, "80th": 1897, "90th": 3579, "95th": 6754, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103264, "hashes": {}}, "samples": 11229, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 18711293, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108647, "hashes": {}}, "samples": 10699, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 19179223, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 25139030, "hashes": {}}, "samples": 4431, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7244372, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 39417140, "total_tokens_skipped": 149, "percentiles": {"0th": 83, "10th": 163, "20th": 257, "30th": 371, "40th": 472, "50th": 647, "60th": 914, "70th": 1326, "80th": 2084, "90th": 4343, "95th": 7170, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00038-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67090304, "hashes": {}}, "samples": 11183, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 18379362, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108382, "hashes": {}}, "samples": 11462, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 18253490, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 29556826, "hashes": {}}, "samples": 4739, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8383398, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00038-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 40501908, "total_tokens_skipped": 0, "percentiles": {"0th": 86, "10th": 164, "20th": 261, "30th": 373, "40th": 470, "50th": 638, "60th": 912, "70th": 1312, "80th": 2068, "90th": 4219, "95th": 7007, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00038-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00039-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab53d3757196a0f38db1ac92553abe5cc4efd6e7a94da6e0fcc414bb1277a2ba
3
+ size 27297616
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00040-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e841c979500d106232451b11d51a43decdd20efeee486d53f0299acba574a20e
3
+ size 49985546
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00044-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ffb37240550f5377b9bf58e86be24fab9d0e188e87437acd4b8d1a204c3b612
3
+ size 67076524
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00044-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e885bd28dd27e52dbb57ddb360a51b43d6c4e272868ff8d568e98a864e4b89a1
3
+ size 67107182
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b661a1bc349815b05e20a410fe631fcdb59ebabe0d3bbad48e7239793837df8b
3
+ size 67103424
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fc27d22c31405e352a8027174f86aa95bcc3f6e81de630b71e19db46e6ce501
3
+ size 67097222
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:663d42aa9ebf11558049ece8ac2e7b3542358903440869fe23f0fe2510660486
3
+ size 67105947
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dee32293e74ac79cdc25358cf6c1baff84dcde9f1baf0c299a9d3877cb7a08e
3
+ size 67091979
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cc30b6ad1526e7748d9bcd3e0133c6bcefc2f71cd3d0123e753f8e9016b0244
3
+ size 67105027
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00052-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84d5bfadf0e1a7aeac8efc5786237a6c8de0ac85e2b6ca382e484ea1cb1cce65
3
+ size 67108618
train/fineweb2-sampled-ext-v2/hun_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00052-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef415fc084c4ce9a09e11585db06f5ed4e7f68453a18ba002eacbd68e3170f6b
3
+ size 67090028