Datasets:

ArXiv:
License:
orionweller commited on
Commit
7b46c03
·
verified ·
1 Parent(s): 7c9b61a

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-decay-v2/epo_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  2. train/fineweb2-sampled-decay-v2/epo_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  3. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0016-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  4. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0016-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  5. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0016-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  6. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0125-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  7. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0202-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  8. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0202-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  9. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0202-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  10. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0235-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  11. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0235-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  12. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0235-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  13. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0369-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  14. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0391-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  15. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0391-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  16. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0391-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  17. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0591-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  18. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0591-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  19. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0591-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  20. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0592-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  21. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0592-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  22. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0592-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  23. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0697-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  24. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0697-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  25. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0697-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  26. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0783-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  27. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0783-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  28. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0783-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  29. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0834-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  30. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0834-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  31. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0834-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  32. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0841-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  33. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0932-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  34. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  36. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  37. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  38. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1137-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  39. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1137-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  40. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1137-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  41. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1171-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  42. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1201-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  43. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1206-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  44. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1402-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  45. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1435-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  46. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1435-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  47. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1435-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  48. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1438-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  49. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1438-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  50. train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1438-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
train/fineweb2-sampled-decay-v2/epo_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26df6ea6d3fd6ec671db671211da8f32ca6954cebdeab5754ef4fd5da7927dd3
3
+ size 67090157
train/fineweb2-sampled-decay-v2/epo_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc6b03e10589cc80d8aa5039b69084e2c904fbedd849383eeec418c579146ed8
3
+ size 67102222
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0016-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105661, "hashes": {}}, "samples": 15221, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 17370682, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22051653, "hashes": {}}, "samples": 4954, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5841371, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0016-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21967012, "total_tokens_skipped": 36, "percentiles": {"0th": 69, "10th": 239, "20th": 341, "30th": 449, "40th": 568, "50th": 705, "60th": 891, "70th": 1142, "80th": 1541, "90th": 2264, "95th": 3175, "99th": 7455, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0016-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0125-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d44b969b5a91fdd1d33b2311991d87320b6862210c7b0d9c4b8b234e1a21fef4
3
+ size 24708402
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0202-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107512, "hashes": {}}, "samples": 14167, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14774182, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 29615999, "hashes": {}}, "samples": 6045, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6287951, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0202-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 23858053, "total_tokens_skipped": 0, "percentiles": {"0th": 63, "10th": 224, "20th": 315, "30th": 441, "40th": 578, "50th": 747, "60th": 972, "70th": 1277, "80th": 1752, "90th": 2496, "95th": 3613, "99th": 8028, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0202-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0235-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67098477, "hashes": {}}, "samples": 14528, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15083131, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 26714831, "hashes": {}}, "samples": 5655, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6114728, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0235-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 23130920, "total_tokens_skipped": 0, "percentiles": {"0th": 71, "10th": 229, "20th": 318, "30th": 437, "40th": 583, "50th": 736, "60th": 936, "70th": 1209, "80th": 1643, "90th": 2428, "95th": 3567, "99th": 7334, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0235-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0369-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6102593ca9fd1c3bd17cbdfaa36b1b69dc03d51b4bebce3ff99928e0b5dc3db4
3
+ size 25718475
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0391-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67096089, "hashes": {}}, "samples": 12401, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15511659, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 30718440, "hashes": {}}, "samples": 7867, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6788516, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0391-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 24129896, "total_tokens_skipped": 3, "percentiles": {"0th": 68, "10th": 225, "20th": 332, "30th": 451, "40th": 603, "50th": 785, "60th": 1004, "70th": 1281, "80th": 1652, "90th": 2422, "95th": 3594, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0391-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0591-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104936, "hashes": {}}, "samples": 14110, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15144621, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 32117895, "hashes": {}}, "samples": 6116, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6147937, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0591-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 24482534, "total_tokens_skipped": 11, "percentiles": {"0th": 65, "10th": 245, "20th": 346, "30th": 486, "40th": 640, "50th": 829, "60th": 1049, "70th": 1314, "80th": 1668, "90th": 2459, "95th": 3584, "99th": 8158, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0591-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0592-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107698, "hashes": {}}, "samples": 10565, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13762677, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 43676675, "hashes": {}}, "samples": 9877, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8811723, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0592-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 27369608, "total_tokens_skipped": 0, "percentiles": {"0th": 68, "10th": 218, "20th": 323, "30th": 442, "40th": 600, "50th": 793, "60th": 1039, "70th": 1362, "80th": 1861, "90th": 3075, "95th": 4618, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0592-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0697-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105249, "hashes": {}}, "samples": 15632, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 19348777, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22308122, "hashes": {}}, "samples": 4545, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6197013, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0697-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 22031015, "total_tokens_skipped": 0, "percentiles": {"0th": 63, "10th": 236, "20th": 335, "30th": 443, "40th": 570, "50th": 709, "60th": 891, "70th": 1150, "80th": 1548, "90th": 2249, "95th": 3210, "99th": 7378, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0697-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0783-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106430, "hashes": {}}, "samples": 15155, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13418502, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21995610, "hashes": {}}, "samples": 5118, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4365523, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0783-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21951892, "total_tokens_skipped": 0, "percentiles": {"0th": 64, "10th": 204, "20th": 284, "30th": 378, "40th": 480, "50th": 637, "60th": 841, "70th": 1097, "80th": 1476, "90th": 2351, "95th": 3565, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0783-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0834-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105604, "hashes": {}}, "samples": 15737, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14175734, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18266754, "hashes": {}}, "samples": 4420, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3890648, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0834-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21021111, "total_tokens_skipped": 50, "percentiles": {"0th": 71, "10th": 229, "20th": 330, "30th": 434, "40th": 552, "50th": 693, "60th": 865, "70th": 1095, "80th": 1465, "90th": 2129, "95th": 3060, "99th": 6826, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0834-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0841-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42ccf77711211c98bd55e28b3553936213978cb526360dea1f167d8b036deec5
3
+ size 19505933
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_0932-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 20882099, "total_tokens_skipped": 0, "percentiles": {"0th": 61, "10th": 236, "20th": 338, "30th": 443, "40th": 551, "50th": 670, "60th": 838, "70th": 1058, "80th": 1406, "90th": 2126, "95th": 3125, "99th": 6913, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106507, "hashes": {}}, "samples": 16838, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13938555, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13148929, "hashes": {}}, "samples": 3291, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2807085, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bbe7ec6cedb13cc6ee49983d767ea2a2116e6febbb6ea90f2e3369b9e9b0349
3
+ size 13148929
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 19742476, "total_tokens_skipped": 43, "percentiles": {"0th": 70, "10th": 209, "20th": 293, "30th": 389, "40th": 490, "50th": 627, "60th": 810, "70th": 1058, "80th": 1381, "90th": 1973, "95th": 2905, "99th": 6559, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1128-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1137-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103716, "hashes": {}}, "samples": 14490, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15446337, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 25836358, "hashes": {}}, "samples": 5697, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6006255, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1137-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 22912569, "total_tokens_skipped": 13, "percentiles": {"0th": 71, "10th": 226, "20th": 327, "30th": 448, "40th": 580, "50th": 736, "60th": 945, "70th": 1233, "80th": 1660, "90th": 2368, "95th": 3278, "99th": 7761, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1137-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1171-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b15ac33d38cd01b9fcbd4b537feff34ad3325a7dc8b530f838e0ac636d8e802b
3
+ size 21511116
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1201-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:960f6de50f9b4966fa8a1ac3fee0f56343899eff22c6b180045119ce9fec9975
3
+ size 23845411
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1206-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1402-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e3889de3793caae8ee08f2a1ba8e7f32484c35042cce299dc7d980386380a92
3
+ size 30240626
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1435-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101195, "hashes": {}}, "samples": 14867, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14871574, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 24020784, "hashes": {}}, "samples": 5324, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5347608, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1435-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 22457900, "total_tokens_skipped": 60, "percentiles": {"0th": 69, "10th": 243, "20th": 339, "30th": 451, "40th": 568, "50th": 707, "60th": 883, "70th": 1138, "80th": 1542, "90th": 2337, "95th": 3389, "99th": 7606, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1435-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1438-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107544, "hashes": {}}, "samples": 14103, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 15772711, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 35662728, "hashes": {}}, "samples": 6127, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6717637, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1438-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 25369269, "total_tokens_skipped": 66, "percentiles": {"0th": 72, "10th": 259, "20th": 369, "30th": 521, "40th": 680, "50th": 865, "60th": 1097, "70th": 1352, "80th": 1739, "90th": 2569, "95th": 3700, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/kor_Hang_train-sampled/batch_1438-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff