Datasets:

ArXiv:
License:
orionweller commited on
Commit
d196031
·
verified ·
1 Parent(s): ffe2191

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  2. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  3. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  4. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  5. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  6. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00012-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  7. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  8. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  9. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  10. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  11. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  12. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00016-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  13. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  14. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  15. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  16. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  17. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  18. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  19. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  20. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds +3 -0
  21. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  22. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  23. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  24. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  25. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  26. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  27. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  28. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  29. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  30. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00046-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  31. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00046-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  32. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00046-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  33. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  34. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  36. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  37. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00055-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  38. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00058-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  39. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  40. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00064-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  41. train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00065-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  42. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  43. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  44. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  45. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0066-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  46. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0096-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  47. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0113-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  48. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0117-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  49. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0145-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  50. train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0197-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00003-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35bd7b676957c579420f9bf0c042d65caf92abbc1cd6334a05d0508b57b2a0c6
3
+ size 33503885
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00010-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe08c2a9ab18b9ae21cfa38799b2aae0029602be4028589c1e035bd4979c680b
3
+ size 31468137
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107344, "hashes": {}}, "samples": 16279, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10541770, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67099621, "hashes": {}}, "samples": 15006, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10454230, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67106528, "hashes": {}}, "samples": 15486, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11075251, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 26794764, "hashes": {}}, "samples": 6407, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 3293499, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 56179732, "total_tokens_skipped": 141, "percentiles": {"0th": 77, "10th": 138, "20th": 192, "30th": 272, "40th": 342, "50th": 415, "60th": 548, "70th": 899, "80th": 1449, "90th": 2543, "95th": 4399, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00011-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00012-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c0df97da9b132b650a336d74d17c07dfbcbb37c7d804bd5fd8e9b92a3608709
3
+ size 44367923
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108557, "hashes": {}}, "samples": 16043, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 10801838, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102904, "hashes": {}}, "samples": 15415, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10350817, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67089691, "hashes": {}}, "samples": 15619, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 10358151, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 17534357, "hashes": {}}, "samples": 4229, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 2913933, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9e067e3e4f275868d4f7b966426a137ea4ccc28dd6c8b15f31c6a8f1d5745b9
3
+ size 17534357
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 53891389, "total_tokens_skipped": 46, "percentiles": {"0th": 81, "10th": 137, "20th": 193, "30th": 270, "40th": 341, "50th": 413, "60th": 534, "70th": 889, "80th": 1403, "90th": 2518, "95th": 4398, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00013-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f17a5152e4e5c3135ae8cb993d88e2bc90b545f0d62930880713e70b4823936
3
+ size 13953371
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00016-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ccf5e7f1a3eeae0532ff243a7ddced626148f10e2f01d115843bbaa6f85b20e
3
+ size 11226851
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00020-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e5f05211a14743478f69c1faab50837cd1b72d42e34da0780452730fe146088
3
+ size 29446201
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df2502803268e79b721f9faf1c04a4b4dfe16215f335bb1b695174a14eece120
3
+ size 27761589
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67078343, "hashes": {}}, "samples": 15609, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 24857881, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22346792, "hashes": {}}, "samples": 5592, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8537920, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73365235fd79f5b21a35abadacf1a57ac627beda6354e0267d35d664e5ee271a
3
+ size 22346792
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 22018448, "total_tokens_skipped": 28, "percentiles": {"0th": 80, "10th": 137, "20th": 191, "30th": 272, "40th": 350, "50th": 427, "60th": 583, "70th": 895, "80th": 1381, "90th": 2377, "95th": 4154, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00024-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67101509, "hashes": {}}, "samples": 13576, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7055952, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67102473, "hashes": {}}, "samples": 13891, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8489662, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67107597, "hashes": {}}, "samples": 13439, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6756042, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67107017, "hashes": {}}, "samples": 15945, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 6270532, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 67105145, "hashes": {}}, "samples": 15040, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 7757812, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00005.mds", "bytes": 18029565, "hashes": {}}, "samples": 3263, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00005.mds.zstd", "bytes": 2243332, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:634fda840e26a39ce43ccdede350acfef6a448103f0ed391080cb3e38ea2ac9d
3
+ size 18029565
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 87190519, "total_tokens_skipped": 69, "percentiles": {"0th": 76, "10th": 140, "20th": 202, "30th": 287, "40th": 358, "50th": 434, "60th": 627, "70th": 1012, "80th": 1590, "90th": 2861, "95th": 5234, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00029-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00032-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67cfb385ebf0b974eb57d9ec538e9a75c125472293e82c2f1daa7d561dddb147
3
+ size 21911990
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67099389, "hashes": {}}, "samples": 14935, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11724573, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107733, "hashes": {}}, "samples": 17146, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11527553, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 33301226, "hashes": {}}, "samples": 6880, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6430940, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 41256259, "total_tokens_skipped": 184, "percentiles": {"0th": 81, "10th": 140, "20th": 194, "30th": 272, "40th": 347, "50th": 421, "60th": 561, "70th": 898, "80th": 1412, "90th": 2541, "95th": 4357, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00035-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105174, "hashes": {}}, "samples": 15509, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11485528, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108718, "hashes": {}}, "samples": 15197, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11675255, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 64472808, "hashes": {}}, "samples": 15511, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 11814727, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00037-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 48935195, "total_tokens_skipped": 148, "percentiles": {"0th": 83, "10th": 141, "20th": 199, "30th": 279, "40th": 351, "50th": 424, "60th": 561, "70th": 908, "80th": 1435, "90th": 2529, "95th": 4407, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00041-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f9ef1bd6b799f5eddedba75c47f838b28b36e50ed624d4ef351c0a6d8206415
3
+ size 37904007
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00046-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108599, "hashes": {}}, "samples": 14750, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7943469, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67095765, "hashes": {}}, "samples": 12957, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8920902, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67098763, "hashes": {}}, "samples": 16163, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 6514866, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 67108477, "hashes": {}}, "samples": 16273, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 7678148, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00004.mds", "bytes": 51934037, "hashes": {}}, "samples": 9959, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00004.mds.zstd", "bytes": 6819676, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00046-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 78969084, "total_tokens_skipped": 449, "percentiles": {"0th": 73, "10th": 142, "20th": 205, "30th": 286, "40th": 360, "50th": 437, "60th": 626, "70th": 999, "80th": 1555, "90th": 2743, "95th": 4943, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00046-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00048-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eca00e7f59042b60ee216c02763f7925eb7a14a1b9ff67fcc3b8d727f8090181
3
+ size 1718617
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108182, "hashes": {}}, "samples": 13715, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9980067, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105610, "hashes": {}}, "samples": 15534, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8346863, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 67108038, "hashes": {}}, "samples": 16249, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 9465613, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00003.mds", "bytes": 56437078, "hashes": {}}, "samples": 12115, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00003.mds.zstd", "bytes": 8609345, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 63521557, "total_tokens_skipped": 12, "percentiles": {"0th": 82, "10th": 142, "20th": 203, "30th": 281, "40th": 353, "50th": 425, "60th": 588, "70th": 938, "80th": 1476, "90th": 2703, "95th": 4936, "99th": 8191, "100th": 8191}}
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00051-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00055-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8b91f8cc942882d4b2cb24351daf5c691a3fda4efd1f810f292515879d47b92
3
+ size 47041846
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00058-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4770a70f242faf0389460cf8a23a3df88b7d8a3cd26abaa9399e009c0ab68eef
3
+ size 8214262
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2cac61b27ac2cdbb69c36f5aa125638ca5160e0939229938376fabd43a9ef55
3
+ size 43345949
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00064-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e39bd0612805fce9e039552ae4d2ec774003f23af48d8e855b5eae57f34b485f
3
+ size 54701481
train/fineweb2-sampled-decay-v2/arb_Arab-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00065-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34df2f587936a5466589330d027d9b20aa01d05d4cfd3df0ecbbdc0386300075
3
+ size 2608964
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8aec49d2b6c51f09425ce6c862dc4fec1cae92ee7017fa125c806150b183a551
3
+ size 67106878
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0013-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31d03c1eefe46d04c48362c96a13f4781bb09edd2d5b90813ba3aab3908b953e
3
+ size 67107777
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0041-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bbeffcccf489dc378f1ed9ff791de630440a6a082a88081ded5ef4713420d62
3
+ size 67106527
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0066-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd62a726faa5c3cae3a1e1116fa3e6bfd9c55d4f81a074f7fdb7ad159e0ed1cd
3
+ size 67107554
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0096-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d601e9f4c8d2c06f54be7a694345b55c94921fb544c4e3ac293f6f68b7b55f3
3
+ size 67101023
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0113-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3deb392df1ee916a128f7c0593c629bd2c757334bf715a3a04e8c94be1c8bf64
3
+ size 67107730
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0117-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:923f760502758bebe589cab4c4abed6013d38ed1ac00499da82db45c91733dc7
3
+ size 67079433
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0145-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52871445a5c34d6b09b8b7b15083662ca36bafe0cb6d523182739578b8c6cf6d
3
+ size 67108185
train/fineweb2-sampled-decay-v2/ekk_Latn_train-sampled/batch_0197-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab95eca0f87c46401f1fc91f4c30b2f4c78d278365c9db1ec9a772cdc43348d6
3
+ size 67108148