Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0010-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
- train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
- train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0018-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
- train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0018-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
- train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0020-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0023-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00007.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
- train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0010-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e461385912c0f80a0580f672cfc52f8ddf68e38fea11221256e8888c1f1b5eb
|
| 3 |
+
size 48946654
|
train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:81c17b2b64719f0787b3d555827d3ba5f549cf9c1da07bb418231da5779be902
|
| 3 |
+
size 44754493
|
train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0018-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cc23460ee60eb1db5e4e9de7d6335713569d6ff66d53734dbe27425ae69d67a6
|
| 3 |
+
size 67080747
|
train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0018-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:553b0c0b1470901829132ef3283fb0f3f12bddd7c66028e9acc0fb3a78c64022
|
| 3 |
+
size 20805896
|
train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0020-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:782ca3af17c998bcaa9eda9e4c7dd0ac3b5547d0beb3ba6196ae7a86eacaec65
|
| 3 |
+
size 45625876
|
train/fineweb2-sampled-decay-v2/ckb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0023-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:32bb08c67d762bd5b1dd393c285f974b5de8cb5c1331ed46ebe5f40295b45705
|
| 3 |
+
size 61569619
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102518, "hashes": {}}, "samples": 12746, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 4728638, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67099103, "hashes": {}}, "samples": 13080, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4875966, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67105627, "hashes": {}}, "samples": 13910, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5076236, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107136, "hashes": {}}, "samples": 11408, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5184608, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67078223, "hashes": {}}, "samples": 12228, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 5574474, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 67108432, "hashes": {}}, "samples": 12125, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 6190186, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00006.mds", "bytes": 67108276, "hashes": {}}, "samples": 13002, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00006.mds.zstd", "bytes": 6612757, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00007.mds", "bytes": 26721372, "hashes": {}}, "samples": 2528, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00007.mds.zstd", "bytes": 1988015, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00007.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c36d5a78757a078077d1d7c07d9c69f0b415bc8bf504bd8a0462e2ae1c3044c6
|
| 3 |
+
size 26721372
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 122655516, "total_tokens_skipped": 223, "percentiles": {"0th": 80, "10th": 170, "20th": 248, "30th": 365, "40th": 469, "50th": 598, "60th": 844, "70th": 1184, "80th": 1820, "90th": 3427, "95th": 6267, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00001-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106876, "hashes": {}}, "samples": 14976, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 5883163, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67093516, "hashes": {}}, "samples": 12544, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7255952, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67105056, "hashes": {}}, "samples": 13400, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7297432, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67105304, "hashes": {}}, "samples": 14805, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6162835, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67106911, "hashes": {}}, "samples": 12097, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 7863403, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 38984301, "hashes": {}}, "samples": 6701, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 3901284, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 92436924, "total_tokens_skipped": 84, "percentiles": {"0th": 80, "10th": 167, "20th": 238, "30th": 341, "40th": 450, "50th": 559, "60th": 765, "70th": 1092, "80th": 1653, "90th": 3021, "95th": 5419, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00006-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67085991, "hashes": {}}, "samples": 12805, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9794187, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67079488, "hashes": {}}, "samples": 12312, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9037238, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67103004, "hashes": {}}, "samples": 12391, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9251465, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107305, "hashes": {}}, "samples": 12394, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 9306040, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 20275224, "hashes": {}}, "samples": 3267, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 3116563, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:15eaebb702080f0ff0b4601e23b13eb45953b68884d9696de6fb9f9597d72307
|
| 3 |
+
size 20275224
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 71314530, "total_tokens_skipped": 55, "percentiles": {"0th": 95, "10th": 173, "20th": 249, "30th": 360, "40th": 469, "50th": 588, "60th": 820, "70th": 1177, "80th": 1770, "90th": 3336, "95th": 6679, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102123, "hashes": {}}, "samples": 11100, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13913895, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67095117, "hashes": {}}, "samples": 11519, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13924246, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107157, "hashes": {}}, "samples": 10737, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 13867864, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 1469034, "hashes": {}}, "samples": 231, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 309371, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:814b8c556ff8111a7636a72eecb0477aae9714da97f71e7244587c537aebeeee
|
| 3 |
+
size 1469034
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 50157397, "total_tokens_skipped": 142, "percentiles": {"0th": 99, "10th": 177, "20th": 260, "30th": 386, "40th": 497, "50th": 651, "60th": 932, "70th": 1309, "80th": 1997, "90th": 4009, "95th": 8190, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67100277, "hashes": {}}, "samples": 12380, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7044892, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67091346, "hashes": {}}, "samples": 11692, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6157455, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67106758, "hashes": {}}, "samples": 15446, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5300343, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107607, "hashes": {}}, "samples": 13631, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 5808513, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67079230, "hashes": {}}, "samples": 11881, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 6968536, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 67108612, "hashes": {}}, "samples": 11214, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 6752643, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00006.mds", "bytes": 25480866, "hashes": {}}, "samples": 4052, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00006.mds.zstd", "bytes": 2391824, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 105737500, "total_tokens_skipped": 40, "percentiles": {"0th": 95, "10th": 173, "20th": 260, "30th": 376, "40th": 481, "50th": 608, "60th": 856, "70th": 1183, "80th": 1802, "90th": 3220, "95th": 5718, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00017-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67086408, "hashes": {}}, "samples": 11902, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8141159, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67098698, "hashes": {}}, "samples": 15255, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6422804, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108788, "hashes": {}}, "samples": 12239, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8077966, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67092959, "hashes": {}}, "samples": 14817, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6459281, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67100556, "hashes": {}}, "samples": 12585, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 7610369, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 13199954, "hashes": {}}, "samples": 2501, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 1645907, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c23f5bd55460b59c32eb795f07498cfb99fb441bc645948e9d860f4774d82c1
|
| 3 |
+
size 13199954
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 86066447, "total_tokens_skipped": 111, "percentiles": {"0th": 84, "10th": 169, "20th": 242, "30th": 348, "40th": 453, "50th": 565, "60th": 772, "70th": 1098, "80th": 1668, "90th": 2975, "95th": 5388, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106925, "hashes": {}}, "samples": 13990, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6750483, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67097194, "hashes": {}}, "samples": 12744, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7484697, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108687, "hashes": {}}, "samples": 14687, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 7377872, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67098722, "hashes": {}}, "samples": 12604, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7406422, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67108748, "hashes": {}}, "samples": 12920, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 7526431, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 11956716, "hashes": {}}, "samples": 3085, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 1032778, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:843640e315fc3bb9af8bdadb8775d3edbd80071b246ebaee63ee59f6bc3d9e62
|
| 3 |
+
size 11956716
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 85752302, "total_tokens_skipped": 263, "percentiles": {"0th": 87, "10th": 167, "20th": 238, "30th": 339, "40th": 445, "50th": 552, "60th": 754, "70th": 1080, "80th": 1635, "90th": 2999, "95th": 5208, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00023-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106316, "hashes": {}}, "samples": 11793, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9729383, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67099438, "hashes": {}}, "samples": 12074, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10122160, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67102626, "hashes": {}}, "samples": 13939, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9773325, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67103684, "hashes": {}}, "samples": 12663, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 10221285, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 213954, "hashes": {}}, "samples": 36, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 44825, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:219b4dd2a889728f993cc5b07d35e328608af6475eecd448e075fc3a466c28a6
|
| 3 |
+
size 213954
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 66350841, "total_tokens_skipped": 42, "percentiles": {"0th": 93, "10th": 170, "20th": 244, "30th": 354, "40th": 463, "50th": 578, "60th": 794, "70th": 1118, "80th": 1709, "90th": 3277, "95th": 6412, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00025-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67085669, "hashes": {}}, "samples": 12290, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9735009, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67095291, "hashes": {}}, "samples": 12229, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9720752, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67082180, "hashes": {}}, "samples": 12116, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9799321, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67106739, "hashes": {}}, "samples": 12814, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 9398594, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 9455212, "hashes": {}}, "samples": 2139, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 1320595, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e4f86eac461a63bdd4506031253603d9b4e7f2e59621407dd600bdb15aecb4a
|
| 3 |
+
size 9455212
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 68633330, "total_tokens_skipped": 115, "percentiles": {"0th": 98, "10th": 170, "20th": 244, "30th": 359, "40th": 459, "50th": 575, "60th": 800, "70th": 1135, "80th": 1741, "90th": 3354, "95th": 6573, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106317, "hashes": {}}, "samples": 12983, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10208666, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67103212, "hashes": {}}, "samples": 11836, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10836906, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107002, "hashes": {}}, "samples": 12382, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 10954914, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 48509200, "hashes": {}}, "samples": 9397, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7820558, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 61713122, "total_tokens_skipped": 0, "percentiles": {"0th": 93, "10th": 170, "20th": 245, "30th": 355, "40th": 469, "50th": 588, "60th": 807, "70th": 1145, "80th": 1737, "90th": 3296, "95th": 6446, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00026-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106915, "hashes": {}}, "samples": 10761, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15158665, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107684, "hashes": {}}, "samples": 10870, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15473762, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 57322314, "hashes": {}}, "samples": 8883, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 12808482, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 47397301, "total_tokens_skipped": 115, "percentiles": {"0th": 95, "10th": 176, "20th": 266, "30th": 389, "40th": 510, "50th": 659, "60th": 943, "70th": 1343, "80th": 2095, "90th": 4464, "95th": 8190, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107431, "hashes": {}}, "samples": 10541, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15265500, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67097920, "hashes": {}}, "samples": 10746, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15117242, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 52073670, "hashes": {}}, "samples": 8552, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 12073840, "hashes": {}}}], "version": 2}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e372bd14a69979f4ad43169b9bf4c3153b1cf650442bd896e17ff0d2566ec302
|
| 3 |
+
size 52073670
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 46093604, "total_tokens_skipped": 52, "percentiles": {"0th": 96, "10th": 177, "20th": 264, "30th": 389, "40th": 506, "50th": 668, "60th": 937, "70th": 1346, "80th": 2122, "90th": 4382, "95th": 8190, "99th": 8191, "100th": 8191}}
|
train/fineweb2-sampled-decay-v2/ell_Grek-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00031-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|