Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/fineweb2-sampled-ext-v2/bel_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/bel_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0036-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/bel_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00044-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00163-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00169-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00319-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00322-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00411-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00411-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00411-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00530-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00530-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00577-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00581-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00756-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00802-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00824-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00908-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00055-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00257-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00257-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00349-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00349-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00457-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00457-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00183-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00183-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00302-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00302-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00302-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00306-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00376-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
train/fineweb2-sampled-ext-v2/bel_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d77640322c0808cf462c35d044ef86f8fd93a9f42da3b98a1628c3fef5db5716
|
3 |
+
size 67106212
|
train/fineweb2-sampled-ext-v2/bel_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0036-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:018bc621dd2203b1ee868966cef1d48da156e6d942ee887e57463647e8ce30e4
|
3 |
+
size 67108085
|
train/fineweb2-sampled-ext-v2/bel_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0090-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:365025db8841bce70450ee89be4218f818e4da3e25b0cb780fa0a8d60aed8174
|
3 |
+
size 67108514
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00044-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e2bf517959a21a376daaa98c7294a5f17f5f270f87bf9ea3e33c137eaf0d58e
|
3 |
+
size 67102508
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00163-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e687989dd36eed01be99eeba795f2e853c33d73d90063e5ff1881bfd59c224fc
|
3 |
+
size 40148805
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00169-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc6c819d2dc7a4a485a64f18b1f26fdc6bb7302ff454a180e011de462d712156
|
3 |
+
size 67097883
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bed6f7592d76bfb572fa8385a9b4caabb5eb96f33ec267888a3e040cd4d920b6
|
3 |
+
size 67108833
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00182-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:76b2008e7593117f2920c75ab91cf05c1974bfdaa88f2d772d2b1c4a669cef1b
|
3 |
+
size 67108072
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00319-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:16585c1aa546bea964cf170fdecf0c0cc68f2f5d397d822c252eddf7f93310a8
|
3 |
+
size 22918521
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00322-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0b974793aa5eaeeb0d0cc7dd461b2b41ca952c001de3e970e6e67ff724dbd996
|
3 |
+
size 19136888
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00411-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67100123, "hashes": {}}, "samples": 17244, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8721695, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102232, "hashes": {}}, "samples": 16592, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9180639, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108207, "hashes": {}}, "samples": 15032, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9038476, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107884, "hashes": {}}, "samples": 16923, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 9410119, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 22709964, "hashes": {}}, "samples": 5455, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 3142381, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00411-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 71645152, "total_tokens_skipped": 170, "percentiles": {"0th": 59, "10th": 190, "20th": 268, "30th": 337, "40th": 410, "50th": 544, "60th": 716, "70th": 961, "80th": 1395, "90th": 2288, "95th": 3412, "99th": 8159, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00411-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00530-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 39586825, "total_tokens_skipped": 156, "percentiles": {"0th": 61, "10th": 203, "20th": 279, "30th": 341, "40th": 409, "50th": 519, "60th": 670, "70th": 858, "80th": 1187, "90th": 1952, "95th": 2920, "99th": 6729, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00530-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00577-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc4620223931bbb02b3dab2506a3d826cc7a05ec0a0545c2ef2cbcaf8bac9aa0
|
3 |
+
size 67108166
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00581-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:285484db546d2d91b33b5110b88b4644241f0728f6e78e382929067a6611fbba
|
3 |
+
size 67105829
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00756-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9f93382dd60b086ebb9a9cd5478b934ad70eae64e332c6ee77e11d98cf220527
|
3 |
+
size 67106820
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00802-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103657, "hashes": {}}, "samples": 16447, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16287621, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107173, "hashes": {}}, "samples": 16154, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16859976, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 24705442, "hashes": {}}, "samples": 6058, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6445392, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00824-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2a244a23830e5292e90315c77cff9c8321afa063617c8a42a38f6dd056aa9b5
|
3 |
+
size 67107345
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00908-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd8c81f995a55ea221b20e91f8a59924e4bd21d6830eb594cf558b40c18629ca
|
3 |
+
size 67107597
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107370, "hashes": {}}, "samples": 14220, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5732651, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107774, "hashes": {}}, "samples": 17632, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5572154, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108566, "hashes": {}}, "samples": 13601, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6098216, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67104303, "hashes": {}}, "samples": 16719, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5277155, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67107376, "hashes": {}}, "samples": 15718, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5880228, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 19393541, "hashes": {}}, "samples": 5361, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 1546607, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 87403628, "total_tokens_skipped": 0, "percentiles": {"0th": 75, "10th": 186, "20th": 275, "30th": 388, "40th": 517, "50th": 664, "60th": 838, "70th": 1062, "80th": 1396, "90th": 2138, "95th": 3308, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67095532, "hashes": {}}, "samples": 16535, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5658587, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108789, "hashes": {}}, "samples": 13319, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5925684, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108837, "hashes": {}}, "samples": 15212, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5734525, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67106079, "hashes": {}}, "samples": 16774, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5769100, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67106565, "hashes": {}}, "samples": 13993, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5901800, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 23235760, "hashes": {}}, "samples": 5282, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 1803994, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105743, "hashes": {}}, "samples": 15836, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6203818, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67098825, "hashes": {}}, "samples": 14700, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6207983, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67103672, "hashes": {}}, "samples": 15683, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6364079, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108201, "hashes": {}}, "samples": 15285, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6193348, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 59177675, "hashes": {}}, "samples": 13888, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5470896, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 80695357, "total_tokens_skipped": 125, "percentiles": {"0th": 77, "10th": 189, "20th": 275, "30th": 391, "40th": 514, "50th": 664, "60th": 843, "70th": 1076, "80th": 1421, "90th": 2180, "95th": 3478, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0011-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107884, "hashes": {}}, "samples": 13301, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16829510, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 57544001, "hashes": {}}, "samples": 11102, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15021863, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 30773364, "total_tokens_skipped": 88, "percentiles": {"0th": 78, "10th": 196, "20th": 309, "30th": 445, "40th": 599, "50th": 757, "60th": 958, "70th": 1236, "80th": 1689, "90th": 2817, "95th": 4444, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0033-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104510, "hashes": {}}, "samples": 13025, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 18210361, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 49587007, "hashes": {}}, "samples": 9342, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14076925, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 28815729, "total_tokens_skipped": 18, "percentiles": {"0th": 77, "10th": 203, "20th": 317, "30th": 456, "40th": 614, "50th": 768, "60th": 972, "70th": 1262, "80th": 1726, "90th": 2898, "95th": 4584, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/cym_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10fa46d708c87de7ceafc3d5dffb0921d79ff6c64e294cdd8f153d91c020d238
|
3 |
+
size 67098633
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80adf451a32672a38ad76e52562b8cafaa77e975e9c78280f1a6d2489b9eeba9
|
3 |
+
size 67107739
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00055-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b50b0c8d6bf8cef3f5441a9832e3b1e31336b32e63621cc6b97b17623adc63ec
|
3 |
+
size 67107147
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00257-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1878e2e5e0f35222e60092a5cc1f33102d241ecc8db6c96a27ede0eb9ec4e961
|
3 |
+
size 67097077
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00257-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e4034bf12851aaaee510215b9c8e22999581da98fdb17f1c15e75e20e458d4
|
3 |
+
size 67108425
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00349-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b9df69aa15d090eaad2066b2fd695254a77b8b391d021fdc9f00849abf69715d
|
3 |
+
size 67103601
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00349-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de806ff2e33f637ff5c2bc8a599b80ddf4d3814bd6a16262899ab47054b558ef
|
3 |
+
size 67100886
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00457-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:071b980add48707a275a027697a2a22bb902794ae9a4688b374f033b1524b855
|
3 |
+
size 67099284
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00457-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da9c5e07c138bc36f9a04250a20b8292a440f06b87a9e000e1cc6e7ba22e6226
|
3 |
+
size 67108160
|
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00183-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103436, "hashes": {}}, "samples": 16866, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12999051, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22071888, "hashes": {}}, "samples": 5905, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4415759, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00183-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00302-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106558, "hashes": {}}, "samples": 16436, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14947186, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107752, "hashes": {}}, "samples": 16833, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16358824, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 20250951, "hashes": {}}, "samples": 5358, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 4396496, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00302-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 38001028, "total_tokens_skipped": 76, "percentiles": {"0th": 59, "10th": 105, "20th": 167, "30th": 255, "40th": 337, "50th": 462, "60th": 656, "70th": 904, "80th": 1301, "90th": 2197, "95th": 3793, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00302-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00306-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6d40a0fb4c1de567c2fe3e3aea63c24e629cb53ff86f1946fda35b4139344f8e
|
3 |
+
size 67104495
|
train/fineweb2-sampled-ext-v2/fra_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00376-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7da44297ceaaa1c0dabb913917564ac068bdbd7bf57ed2fa9d2bd1705fc84579
|
3 |
+
size 52580352
|