Datasets:

ArXiv:
License:
orionweller commited on
Commit
58978e7
·
verified ·
1 Parent(s): 923236b

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0156-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  2. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0437-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  3. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0449-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  4. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0616-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  5. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0795-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  6. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0868-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  7. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0875-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  8. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0981-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  9. train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_1047-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  10. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  11. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  12. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  13. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  14. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  15. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  16. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00081-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  17. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00081-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  18. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00081-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  19. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  20. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  21. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  22. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  23. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00096-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  24. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00096-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  25. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00096-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  26. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  27. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  28. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  29. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  30. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  31. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  32. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  33. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  34. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  35. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  36. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00156-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  37. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00156-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  38. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00156-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  39. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  40. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  41. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  42. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00201-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  43. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00201-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  44. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00201-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  45. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00216-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  46. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00216-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  47. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00216-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  48. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00262-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  49. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00262-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  50. train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00262-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0156-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31f9511dbf6fec560540f5dc886ba6ad9974efdf4062dc261736bd16e821ed4e
3
+ size 67101892
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0437-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4acc5adbfdde0da1b152af18d3f9d9b71e4119380067d22cb936ec2dd3abddf4
3
+ size 67090465
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0449-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b1355858104f7305607721d822943e2e13a4b5356c1bb2ace6e2ede69bea7b5
3
+ size 67106969
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0616-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51c438a94c66737cf74e69e30628ffeb4c2193ab0837b4f695c18e078f415718
3
+ size 67105504
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0795-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c891dd0b92fe8b860631512b41a52e9ecb4e7a651ab3f93b3752d13c5baa00d
3
+ size 67101782
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0868-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f18bf4d12872090b8d46d47c151c1cda68ed2656045c18b82fa02614449c422
3
+ size 67104625
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0875-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af2498cb4dfc3db738cd0a2527fda923ed466d24fe94745323baf52d1c70882e
3
+ size 67106972
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_0981-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdb03d372e75e6fd1c63c49daf8b70a99cb1b9d4dbd43077153dac0cb05876e3
3
+ size 67106303
train/fineweb2-sampled-decay-v2/hin_Deva_train-sampled/batch_1047-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4f43c30af3e98ddbf315b767b2c7b07a1831ea7bb6a5240942368ec522326b3
3
+ size 67107142
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108735, "hashes": {}}, "samples": 27628, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8096283, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108026, "hashes": {}}, "samples": 24802, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7222017, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 65860766, "hashes": {}}, "samples": 30085, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 8035961, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 48708397, "total_tokens_skipped": 95, "percentiles": {"0th": 46, "10th": 131, "20th": 167, "30th": 198, "40th": 234, "50th": 274, "60th": 325, "70th": 421, "80th": 703, "90th": 1260, "95th": 2022, "99th": 6175, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00018-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105757, "hashes": {}}, "samples": 27860, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9283375, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67108367, "hashes": {}}, "samples": 27645, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9510774, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 26732453, "hashes": {}}, "samples": 9766, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3758445, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 39200102, "total_tokens_skipped": 60, "percentiles": {"0th": 43, "10th": 108, "20th": 153, "30th": 185, "40th": 220, "50th": 261, "60th": 314, "70th": 398, "80th": 660, "90th": 1221, "95th": 2100, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00062-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00081-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107283, "hashes": {}}, "samples": 26291, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11577586, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 64643660, "hashes": {}}, "samples": 26143, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11505687, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00081-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 32104722, "total_tokens_skipped": 0, "percentiles": {"0th": 42, "10th": 109, "20th": 156, "30th": 189, "40th": 228, "50th": 275, "60th": 335, "70th": 452, "80th": 780, "90th": 1386, "95th": 2204, "99th": 5814, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00081-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105514, "hashes": {}}, "samples": 26237, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11399563, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107434, "hashes": {}}, "samples": 24989, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11492570, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 7878379, "hashes": {}}, "samples": 3236, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 1620730, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c6f4fc0b3d8501939b6a30df70f131e2d71e313ba445ae2bdf06a0869a5a21
3
+ size 7878379
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 34657516, "total_tokens_skipped": 50, "percentiles": {"0th": 47, "10th": 109, "20th": 156, "30th": 190, "40th": 230, "50th": 276, "60th": 338, "70th": 463, "80th": 787, "90th": 1425, "95th": 2256, "99th": 7391, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00090-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00096-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107578, "hashes": {}}, "samples": 25754, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13361168, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22534502, "hashes": {}}, "samples": 8676, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3253070, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00096-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21863564, "total_tokens_skipped": 0, "percentiles": {"0th": 46, "10th": 109, "20th": 156, "30th": 188, "40th": 225, "50th": 272, "60th": 336, "70th": 457, "80th": 773, "90th": 1365, "95th": 2184, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00096-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67077335, "hashes": {}}, "samples": 24747, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11807743, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67107471, "hashes": {}}, "samples": 24355, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11897707, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 6106505, "hashes": {}}, "samples": 1988, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 1233516, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6af4ebb79350a46981ff3f81949cf7e7051de98ae7f29f0e293f21a86fc08a5
3
+ size 6106505
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 34260759, "total_tokens_skipped": 0, "percentiles": {"0th": 47, "10th": 113, "20th": 161, "30th": 196, "40th": 240, "50th": 295, "60th": 374, "70th": 583, "80th": 924, "90th": 1568, "95th": 2387, "99th": 6068, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00120-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104407, "hashes": {}}, "samples": 25396, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12750982, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 56590136, "hashes": {}}, "samples": 21508, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11288261, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 30178194, "total_tokens_skipped": 37, "percentiles": {"0th": 44, "10th": 112, "20th": 160, "30th": 194, "40th": 238, "50th": 290, "60th": 364, "70th": 555, "80th": 884, "90th": 1512, "95th": 2299, "99th": 5258, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00134-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107289, "hashes": {}}, "samples": 25542, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14398154, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 41846099, "hashes": {}}, "samples": 15875, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9649331, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 26580141, "total_tokens_skipped": 16, "percentiles": {"0th": 43, "10th": 111, "20th": 159, "30th": 192, "40th": 234, "50th": 288, "60th": 361, "70th": 540, "80th": 866, "90th": 1503, "95th": 2329, "99th": 5275, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00140-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00156-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106462, "hashes": {}}, "samples": 27513, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 9103483, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12658987, "hashes": {}}, "samples": 5163, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2041079, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00156-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 19422147, "total_tokens_skipped": 100, "percentiles": {"0th": 48, "10th": 124, "20th": 166, "30th": 199, "40th": 236, "50th": 275, "60th": 328, "70th": 428, "80th": 718, "90th": 1260, "95th": 2077, "99th": 6094, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00156-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107563, "hashes": {}}, "samples": 27854, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8352889, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67100295, "hashes": {}}, "samples": 29100, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8565184, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 40513187, "hashes": {}}, "samples": 16322, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5381575, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 42516273, "total_tokens_skipped": 0, "percentiles": {"0th": 43, "10th": 121, "20th": 163, "30th": 192, "40th": 228, "50th": 269, "60th": 323, "70th": 413, "80th": 692, "90th": 1235, "95th": 1999, "99th": 5888, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00184-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00201-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102785, "hashes": {}}, "samples": 29486, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8826401, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105641, "hashes": {}}, "samples": 28934, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9072720, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 22551408, "hashes": {}}, "samples": 9927, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 3244948, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00201-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 38104152, "total_tokens_skipped": 0, "percentiles": {"0th": 46, "10th": 112, "20th": 159, "30th": 191, "40th": 228, "50th": 274, "60th": 330, "70th": 432, "80th": 740, "90th": 1273, "95th": 1910, "99th": 4523, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00201-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00216-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67105657, "hashes": {}}, "samples": 28677, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8193721, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67105619, "hashes": {}}, "samples": 28693, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8365849, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 41109741, "hashes": {}}, "samples": 17290, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 5312131, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00216-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 42644194, "total_tokens_skipped": 30, "percentiles": {"0th": 48, "10th": 111, "20th": 159, "30th": 192, "40th": 228, "50th": 274, "60th": 330, "70th": 433, "80th": 734, "90th": 1269, "95th": 1979, "99th": 5010, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00216-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00262-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106775, "hashes": {}}, "samples": 25056, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 17355052, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 23022029, "hashes": {}}, "samples": 8497, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6456867, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00262-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 21998833, "total_tokens_skipped": 0, "percentiles": {"0th": 45, "10th": 110, "20th": 161, "30th": 198, "40th": 242, "50th": 296, "60th": 370, "70th": 564, "80th": 897, "90th": 1526, "95th": 2302, "99th": 5624, "100th": 8191}}
train/fineweb2-sampled-decay-v2/jpn_Jpan-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00262-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff