Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0241-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0241-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0516-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0516-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0516-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00111-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00151-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00151-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00306-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00306-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00309-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00315-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00315-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00326-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00326-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00326-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00350-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00350-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00350-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00354-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00365-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00389-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00434-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00525-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00525-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00525-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0017-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0043-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0046-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0050-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0052-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0299-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0299-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0309-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0397-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0679-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0679-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00115-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:946a73ca9bbcf37e730807fb4267acf497de399d55104be4128bed34e9bb1875
|
3 |
+
size 24295505
|
train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0241-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 22687391, "total_tokens_skipped": 92, "percentiles": {"0th": 85, "10th": 171, "20th": 241, "30th": 313, "40th": 400, "50th": 531, "60th": 679, "70th": 899, "80th": 1336, "90th": 2479, "95th": 4653, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0241-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0516-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107772, "hashes": {}}, "samples": 14931, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14387104, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 25433811, "hashes": {}}, "samples": 5550, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5448641, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0516-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 22808458, "total_tokens_skipped": 22, "percentiles": {"0th": 87, "10th": 197, "20th": 282, "30th": 373, "40th": 481, "50th": 616, "60th": 778, "70th": 1025, "80th": 1425, "90th": 2337, "95th": 3976, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/bul_Cyrl_train-sampled/batch_0516-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/cmn_Hani-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00110-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108678, "hashes": {}}, "samples": 15655, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14621622, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102687, "hashes": {}}, "samples": 16327, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13344825, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 65476213, "hashes": {}}, "samples": 16274, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11613802, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:866d57ea27b4ee4cd4b3d2bed2286a135d3d14d120c08d2e7799f51eaf3ecc26
|
3 |
+
size 67106257
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f0b1a6c9fcd7a180edf069ba514ed12eb739d506c6b84ffcf0de341598f43fdf
|
3 |
+
size 67104961
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00111-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 49066181, "total_tokens_skipped": 18, "percentiles": {"0th": 61, "10th": 108, "20th": 158, "30th": 243, "40th": 321, "50th": 421, "60th": 605, "70th": 838, "80th": 1183, "90th": 1930, "95th": 3105, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00151-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de851fe2724a8e14e1344d56f59a64e2859556d47f946680f5f2d7d4fc735359
|
3 |
+
size 67108484
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00151-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b54cfb6b92de6ebf546d0cc7fc279587ed54c03881c926b4848332398b5bfdf8
|
3 |
+
size 67108759
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107589, "hashes": {}}, "samples": 17515, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 18775048, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 38426233, "hashes": {}}, "samples": 9833, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10819222, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00278-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00306-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a76ab22d74488d3efbb90fa23ac45c33879242693532428acdeeeb4ca522daee
|
3 |
+
size 67100081
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00306-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26c6340e0944d46c29374138830080120ec99af8fb829645bf2f74e96fd232bc
|
3 |
+
size 67108460
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00309-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ac0efbdd4311310ad39b18a0707fd8691b932b9d63fc76b7230bc3d0987fb90
|
3 |
+
size 49898013
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00315-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c659e8a722efb4ad974dbcd04f6d615b0a9961176f98f6275f8fe4407f3109fa
|
3 |
+
size 67108631
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00315-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01fc6c7f262603064156c37007d764727462a3d670dbc783ff0696472ddc83af
|
3 |
+
size 67106887
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00326-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105924, "hashes": {}}, "samples": 19709, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8351725, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 61122215, "hashes": {}}, "samples": 20055, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7799230, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00326-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 31423921, "total_tokens_skipped": 0, "percentiles": {"0th": 63, "10th": 108, "20th": 156, "30th": 237, "40th": 310, "50th": 388, "60th": 532, "70th": 743, "80th": 1097, "90th": 1751, "95th": 2737, "99th": 6849, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00326-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00350-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106011, "hashes": {}}, "samples": 19481, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11677321, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67090386, "hashes": {}}, "samples": 18740, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12306016, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 49070893, "hashes": {}}, "samples": 13919, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8997591, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00350-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 44986515, "total_tokens_skipped": 114, "percentiles": {"0th": 62, "10th": 112, "20th": 165, "30th": 244, "40th": 321, "50th": 416, "60th": 580, "70th": 787, "80th": 1107, "90th": 1889, "95th": 3128, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00350-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00354-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:18428de7c5a2339e24cfd334857dbf5a112383286da0744d88e3e66d66bf59e4
|
3 |
+
size 57733793
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00365-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a147aeb04e625bb8254dc996e3a482b269774d12145165f93dbf6c0b30ecebc5
|
3 |
+
size 41799188
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00389-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3fa15259e13516adcb4bd72b70f8edfd616bb6cf3658a268f6b7ec995c47def
|
3 |
+
size 67094265
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102171, "hashes": {}}, "samples": 17180, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11258292, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67106359, "hashes": {}}, "samples": 17686, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10358197, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67097829, "hashes": {}}, "samples": 16719, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11645241, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 22072851, "hashes": {}}, "samples": 6325, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 3828504, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 54922300, "total_tokens_skipped": 70, "percentiles": {"0th": 61, "10th": 110, "20th": 165, "30th": 260, "40th": 339, "50th": 456, "60th": 647, "70th": 895, "80th": 1293, "90th": 2164, "95th": 3563, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00396-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00434-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:23920c745229a2edf38cfbbc348ec77b5acdb509a334c33433209f597c263da2
|
3 |
+
size 67085695
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00525-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107914, "hashes": {}}, "samples": 18984, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9962192, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108147, "hashes": {}}, "samples": 18141, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10662419, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67103513, "hashes": {}}, "samples": 17617, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 10444034, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 24669893, "hashes": {}}, "samples": 7353, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 4385973, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00525-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 55508448, "total_tokens_skipped": 19, "percentiles": {"0th": 61, "10th": 108, "20th": 160, "30th": 247, "40th": 330, "50th": 431, "60th": 614, "70th": 852, "80th": 1213, "90th": 2033, "95th": 3231, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/deu_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00525-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0017-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:414457ffa0e0043025c4a9243b8eb0264436d04473a0040e8622fd6155186540
|
3 |
+
size 41979636
|
train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0043-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49a8ac6121dd30f8cb20c542f6925ad919a5d6ae9c6e0fe5c58e17585b4a4c97
|
3 |
+
size 41059565
|
train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0046-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c3f85276135018129b7c76b5a75c5c173f3d7477ef489933b706e3fb42b2a21
|
3 |
+
size 48779958
|
train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0050-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c8284946db9d8e7a72dd42169e77905a89588f83821444fc77bd89a51dee27ab
|
3 |
+
size 67108084
|
train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd240b84bcfc7b3e23757078566bf100f9d7c127359ccc28c4dfa27bbd478e87
|
3 |
+
size 67108055
|
train/fineweb2-sampled-ext-v2/gle_Latn_train-sampled/batch_0052-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a05a86d3ebd413d235e6b36bd0f7834847e5fa530fb789e8bca22fb984a4cc5e
|
3 |
+
size 46395917
|
train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0299-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101187, "hashes": {}}, "samples": 10477, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12409912, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 47857587, "hashes": {}}, "samples": 9988, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8463664, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0299-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0309-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2d0fb5835135c605de272e5dc72b855c9c1b3e4ad8f9b3bf174e44166c7df72
|
3 |
+
size 67105951
|
train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0397-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86719da5f31279673c9f9d6e59f2900e54f9f046bc09b774ce42c1ad76199bdb
|
3 |
+
size 37246312
|
train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0679-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 24866703, "total_tokens_skipped": 77, "percentiles": {"0th": 95, "10th": 218, "20th": 324, "30th": 461, "40th": 603, "50th": 768, "60th": 975, "70th": 1232, "80th": 1622, "90th": 2546, "95th": 3925, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/heb_Hebr_train-sampled/batch_0679-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 84074177, "total_tokens_skipped": 295, "percentiles": {"0th": 63, "10th": 104, "20th": 142, "30th": 205, "40th": 271, "50th": 336, "60th": 435, "70th": 675, "80th": 1068, "90th": 1990, "95th": 3535, "99th": 8190, "100th": 8191}}
|
train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
train/fineweb2-sampled-ext-v2/ita_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00115-batch_0004-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6039188910ba24eb64c06228d0829d8b41ae7014a239baf78e7854ad31d0c8a4
|
3 |
+
size 14724933
|