Datasets:

ArXiv:
License:
orionweller commited on
Commit
9c09dcf
·
verified ·
1 Parent(s): a961029

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  2. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  3. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  4. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds +3 -0
  5. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  6. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  7. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  8. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  9. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  10. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  11. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  12. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  13. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
  14. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  15. train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  16. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  17. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  18. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  19. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  20. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  21. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  22. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  23. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  24. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  25. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  26. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  27. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  28. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0070-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  29. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0070-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  30. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0070-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  31. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  32. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  33. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  34. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  36. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  37. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0088-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  38. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0088-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  39. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0088-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  40. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0094-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  41. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0094-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  42. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0094-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  43. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0101-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  44. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0101-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  45. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0101-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  46. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0106-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  47. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0106-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  48. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0106-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  49. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0117-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  50. train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0117-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:920ff9b6e8ecc9be1c8be9b809d7dffc2adde22b08a706bc0928228fa7e7bb51
3
+ size 67103980
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f2a4a942c7a04cc3093a7655d06f35a5d8806316b590a6008f732379cf1fcfb
3
+ size 67082218
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ddb16c32ae1d9c3256373839c2ee78a6a01c2343a6fbf6c829aed695f7e3403
3
+ size 67108338
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00002-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a86f608036c8d6d3927ec4e604859bd96b6e48d8c4e8e723714c150b9f2925ca
3
+ size 67082616
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00005-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:552d3efa3c1d1ec7f2f2da25bf0500b7b83ba02f6e906d5677e75a673a778044
3
+ size 67086841
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fde7e5f10522b640f9df2da2df3d3a955bfa24026cb7bd176cee2f63cf4f8d1
3
+ size 67099865
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5a327c84f727d6ea6821b7c93a42b96f0d141b871610b36f11456abe7189050
3
+ size 67108642
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00015-batch_0000-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d3f6988fcbcfcdd97ffc73b42fc17911ff654f28e42053da1a79cd0476f6e79
3
+ size 67107454
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00021-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90f09bc043dad26ef450df470fe9b7964dacc91dd8d61557de0396b10d5e8aa1
3
+ size 67106581
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74ee8e3601e9cb4c491e1777bde23950c37b68939418a85e031da07f061a1d37
3
+ size 67108451
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edebe04e4290f210c6a998687ece2a869803845829bd84d0a90a1591d0f8666b
3
+ size 67100806
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00030-batch_0003-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:859534a730079b1336b8cf557e2743caa8e7a8fb36a5f7762fd7e195e3be1e00
3
+ size 67108637
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17e2e485c4510867d582482e664222b7378640d73ca0830ed475c436b20688a6
3
+ size 67102032
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1685f7516c3a9025010d193e1a68533121edd95e6ba75beaa4df1c4af2e0aff
3
+ size 67096744
train/fineweb2-sampled-decay-v2/dan_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00033-batch_0002-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9aeb05e5f028252914e9b570b009fd435a82509e478d299f1b07a5d8652489a6
3
+ size 67108576
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 53346116, "hashes": {}}, "samples": 20075, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 6982709, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 13016935, "total_tokens_skipped": 0, "percentiles": {"0th": 65, "10th": 126, "20th": 177, "30th": 234, "40th": 309, "50th": 396, "60th": 509, "70th": 635, "80th": 859, "90th": 1312, "95th": 1953, "99th": 5020, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0007-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 57512691, "hashes": {}}, "samples": 20116, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7909055, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 14057785, "total_tokens_skipped": 0, "percentiles": {"0th": 63, "10th": 130, "20th": 191, "30th": 254, "40th": 327, "50th": 420, "60th": 530, "70th": 662, "80th": 880, "90th": 1338, "95th": 2115, "99th": 6137, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0008-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 57713378, "hashes": {}}, "samples": 20136, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7949044, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 14107754, "total_tokens_skipped": 36, "percentiles": {"0th": 67, "10th": 125, "20th": 178, "30th": 238, "40th": 313, "50th": 400, "60th": 516, "70th": 660, "80th": 903, "90th": 1442, "95th": 2222, "99th": 5966, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0026-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 56934507, "hashes": {}}, "samples": 20137, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8138710, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 13912871, "total_tokens_skipped": 0, "percentiles": {"0th": 65, "10th": 136, "20th": 195, "30th": 258, "40th": 330, "50th": 418, "60th": 529, "70th": 668, "80th": 874, "90th": 1344, "95th": 1971, "99th": 6007, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0051-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0070-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 53434039, "hashes": {}}, "samples": 20097, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7716854, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0070-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 13038509, "total_tokens_skipped": 0, "percentiles": {"0th": 70, "10th": 127, "20th": 182, "30th": 241, "40th": 312, "50th": 397, "60th": 505, "70th": 649, "80th": 863, "90th": 1297, "95th": 1868, "99th": 4849, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0070-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 55250974, "hashes": {}}, "samples": 20141, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8134494, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 13491996, "total_tokens_skipped": 0, "percentiles": {"0th": 67, "10th": 126, "20th": 187, "30th": 250, "40th": 320, "50th": 409, "60th": 510, "70th": 641, "80th": 863, "90th": 1331, "95th": 1923, "99th": 5932, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0079-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 51769980, "hashes": {}}, "samples": 20055, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7897040, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 12623067, "total_tokens_skipped": 0, "percentiles": {"0th": 71, "10th": 126, "20th": 185, "30th": 255, "40th": 323, "50th": 407, "60th": 508, "70th": 649, "80th": 857, "90th": 1214, "95th": 1745, "99th": 4483, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0088-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 52204089, "hashes": {}}, "samples": 20052, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7843137, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0088-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 12731670, "total_tokens_skipped": 0, "percentiles": {"0th": 69, "10th": 130, "20th": 186, "30th": 250, "40th": 323, "50th": 407, "60th": 509, "70th": 648, "80th": 838, "90th": 1265, "95th": 1872, "99th": 4140, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0088-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0094-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 53779927, "hashes": {}}, "samples": 20073, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8044862, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0094-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 13125301, "total_tokens_skipped": 65, "percentiles": {"0th": 69, "10th": 128, "20th": 187, "30th": 250, "40th": 319, "50th": 404, "60th": 507, "70th": 634, "80th": 852, "90th": 1279, "95th": 1940, "99th": 5066, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0094-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0101-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 53495501, "hashes": {}}, "samples": 20119, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7973379, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0101-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 13053516, "total_tokens_skipped": 0, "percentiles": {"0th": 66, "10th": 126, "20th": 183, "30th": 242, "40th": 314, "50th": 395, "60th": 505, "70th": 647, "80th": 847, "90th": 1247, "95th": 1894, "99th": 5012, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0101-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0106-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 51720634, "hashes": {}}, "samples": 20064, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 7944817, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0106-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 12610680, "total_tokens_skipped": 0, "percentiles": {"0th": 66, "10th": 125, "20th": 179, "30th": 241, "40th": 313, "50th": 394, "60th": 505, "70th": 641, "80th": 851, "90th": 1266, "95th": 1840, "99th": 4122, "100th": 8191}}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0106-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0117-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 52254663, "hashes": {}}, "samples": 20064, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 8052343, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/glg_Latn_train-sampled/batch_0117-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 12744155, "total_tokens_skipped": 0, "percentiles": {"0th": 66, "10th": 129, "20th": 183, "30th": 248, "40th": 317, "50th": 403, "60th": 507, "70th": 644, "80th": 845, "90th": 1250, "95th": 1844, "99th": 4281, "100th": 8191}}