Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_11-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_11-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_13-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_13-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_13-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_14-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_14-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_14-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_15-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_15-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_15-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_4-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_4-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_4-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_6-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_6-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_6-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_7-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/num_tokens.json +3 -0
- train/multi-wikis/bul_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/multi-wikis/bul_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/multi-wikis/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/multi-wikis/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 30148434, "hashes": {}}, "samples": 50007, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9607748, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7188334, "total_tokens_skipped": 0, "percentiles": {"0th": 28, "10th": 57, "20th": 64, "30th": 71, "40th": 81, "50th": 95, "60th": 109, "70th": 128, "80th": 175, "90th": 280, "95th": 414, "99th": 769, "100th": 8191}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 30577930, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9718551, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7292936, "total_tokens_skipped": 0, "percentiles": {"0th": 27, "10th": 58, "20th": 64, "30th": 71, "40th": 83, "50th": 98, "60th": 112, "70th": 134, "80th": 186, "90th": 287, "95th": 422, "99th": 782, "100th": 7384}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_1-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 29622657, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9584282, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7041654, "total_tokens_skipped": 0, "percentiles": {"0th": 27, "10th": 57, "20th": 62, "30th": 68, "40th": 77, "50th": 91, "60th": 108, "70th": 128, "80th": 176, "90th": 282, "95th": 410, "99th": 777, "100th": 6934}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_10-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_11-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 30682583, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9953169, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_11-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_13-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 28967908, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9114253, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_13-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 6878052, "total_tokens_skipped": 0, "percentiles": {"0th": 27, "10th": 57, "20th": 63, "30th": 69, "40th": 76, "50th": 88, "60th": 107, "70th": 124, "80th": 170, "90th": 271, "95th": 396, "99th": 754, "100th": 6203}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_13-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_14-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 29965550, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9548783, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_14-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7127387, "total_tokens_skipped": 0, "percentiles": {"0th": 24, "10th": 56, "20th": 62, "30th": 70, "40th": 80, "50th": 96, "60th": 111, "70th": 127, "80th": 172, "90th": 278, "95th": 411, "99th": 792, "100th": 8190}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_14-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_15-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 28992531, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9221464, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_15-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 6884193, "total_tokens_skipped": 0, "percentiles": {"0th": 26, "10th": 58, "20th": 62, "30th": 68, "40th": 75, "50th": 87, "60th": 105, "70th": 124, "80th": 170, "90th": 271, "95th": 399, "99th": 763, "100th": 8190}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_15-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 12342934, "hashes": {}}, "samples": 21124, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 3883446, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 2931972, "total_tokens_skipped": 0, "percentiles": {"0th": 28, "10th": 57, "20th": 63, "30th": 70, "40th": 78, "50th": 87, "60th": 102, "70th": 120, "80th": 162, "90th": 265, "95th": 409, "99th": 847, "100th": 4956}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_16-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 29723713, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9200752, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7066980, "total_tokens_skipped": 0, "percentiles": {"0th": 28, "10th": 58, "20th": 66, "30th": 72, "40th": 84, "50th": 100, "60th": 110, "70th": 124, "80th": 168, "90th": 273, "95th": 404, "99th": 763, "100th": 8190}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7099384, "total_tokens_skipped": 0, "percentiles": {"0th": 27, "10th": 58, "20th": 64, "30th": 71, "40th": 84, "50th": 97, "60th": 109, "70th": 125, "80th": 171, "90th": 277, "95th": 409, "99th": 770, "100th": 5520}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_3-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_4-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 30286350, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9758613, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_4-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7207561, "total_tokens_skipped": 0, "percentiles": {"0th": 19, "10th": 57, "20th": 64, "30th": 71, "40th": 82, "50th": 95, "60th": 109, "70th": 134, "80th": 185, "90th": 285, "95th": 414, "99th": 764, "100th": 6258}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_4-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 30609171, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9262617, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7288334, "total_tokens_skipped": 0, "percentiles": {"0th": 20, "10th": 59, "20th": 67, "30th": 76, "40th": 91, "50th": 104, "60th": 113, "70th": 129, "80th": 183, "90th": 276, "95th": 414, "99th": 792, "100th": 4870}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_5-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_6-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 29986298, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9749378, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_6-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7132565, "total_tokens_skipped": 0, "percentiles": {"0th": 27, "10th": 57, "20th": 63, "30th": 70, "40th": 80, "50th": 93, "60th": 108, "70th": 127, "80th": 175, "90th": 282, "95th": 417, "99th": 799, "100th": 7267}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_6-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_7-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 30079030, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9829497, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7155678, "total_tokens_skipped": 0, "percentiles": {"0th": 25, "10th": 57, "20th": 62, "30th": 68, "40th": 78, "50th": 92, "60th": 110, "70th": 131, "80th": 181, "90th": 288, "95th": 414, "99th": 769, "100th": 8190}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_8-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 30628371, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9898832, "hashes": {}}}], "version": 2}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 7293015, "total_tokens_skipped": 0, "percentiles": {"0th": 28, "10th": 58, "20th": 65, "30th": 73, "40th": 82, "50th": 95, "60th": 110, "70th": 129, "80th": 180, "90th": 289, "95th": 425, "99th": 806, "100th": 8190}}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/articles_9-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"version": 2, "shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_2-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 29723713, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_2-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9200752, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_5-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30609171, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_5-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9262617, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_16-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 12342934, "hashes": {}}, "samples": 21124, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_16-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 3883446, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_13-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 28967908, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_13-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9114253, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_10-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 29622657, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_10-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9584282, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_6-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 29986298, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_6-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9749378, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30148434, "hashes": {}}, "samples": 50007, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9607748, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_4-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30286350, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_4-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9758613, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_9-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30628371, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_9-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9898832, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_15-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 28992531, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_15-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9221464, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30577930, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_1-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9718551, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_8-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30079030, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_8-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9829497, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_3-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 29853389, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_3-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9375345, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_7-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30867899, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_7-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9618729, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_11-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30682583, "hashes": {}}, "samples": 50000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_11-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9953169, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_14-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 29965550, "hashes": {}}, "samples": 50001, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_14-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9548783, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "articles_12-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds", "bytes": 30138547, "hashes": {}}, "samples": 50005, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "articles_12-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds.zstd", "bytes": 9681591, "hashes": {}}}]}
|
train/multi-wikis/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups/num_tokens.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"num_tokens": 117418155
|
| 3 |
+
}
|
train/multi-wikis/bul_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6bec5a20c34da61c7bac009bf612e17a17c8e31be2404badc18add67aa836ed
|
| 3 |
+
size 36834765
|
train/multi-wikis/bul_Cyrl-tokenized-chunked-8192-512-32-backfill-nodups/articles_2-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ad3d20999584e2a0eaefbdc7c594b8128450e1a0e7cee06d4f92a3744d31c72
|
| 3 |
+
size 39083000
|
train/multi-wikis/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 56547750, "hashes": {}}, "samples": 50004, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 16347974, "hashes": {}}}], "version": 2}
|
train/multi-wikis/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups/articles_0-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 13785066, "total_tokens_skipped": 0, "percentiles": {"0th": 31, "10th": 77, "20th": 101, "30th": 127, "40th": 158, "50th": 194, "60th": 236, "70th": 298, "80th": 394, "90th": 567, "95th": 765, "99th": 1242, "100th": 8190}}
|