Datasets:

ArXiv:
License:
orionweller commited on
Commit
fc3d460
·
verified ·
1 Parent(s): cf7999f

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train/fineweb2-sampled-decay-v2/gmh_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  2. train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
  3. train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds +3 -0
  4. train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00006.mds +3 -0
  5. train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00007.mds +3 -0
  6. train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00009.mds +3 -0
  7. train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00007.mds +3 -0
  8. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  9. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0058-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  10. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0062-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  11. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  12. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  13. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0100-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  14. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0162-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  15. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0225-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  16. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0226-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  17. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0314-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  18. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0314-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  19. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0314-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  20. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0315-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  21. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0322-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  22. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0353-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  23. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0457-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  24. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0523-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  25. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0551-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  26. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0566-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  27. train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0653-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  28. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  29. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  30. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  31. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0128-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  32. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0128-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  33. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0128-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  34. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  35. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
  36. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  37. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  38. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0264-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  39. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0264-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  40. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0264-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  41. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0356-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  42. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0356-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  43. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0356-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  44. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0372-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  45. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0372-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  46. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0372-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  47. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0403-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
  48. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0403-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
  49. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0403-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
  50. train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0415-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
train/fineweb2-sampled-decay-v2/gmh_Latn-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0006-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce538f7b7c2046637a34983fb4096fdc819deee8bbc2e3989ce504fc68c56a0c
3
+ size 67099872
train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d6cf48445ac5cb2d33c5b996e0eb481419cda454fbca591c2f6c199cb617c0d
3
+ size 67105661
train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:601325264ab950d4dfb78a6065cefad198f18ac08a0959ace529cb8c1cb50977
3
+ size 67100632
train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49eca080ad77fc0eef893b49e11bb65289456e674bb114b35527aa16417f21ea
3
+ size 67108479
train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17faad019d83d4bd8fd86054ce3d2aa155a53bae86809bf5304517fa8044793a
3
+ size 67107680
train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0001-tokenized-chunked-8192-512-32-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9d2eabac1c57909522b71e534ed4373579589b1e040d7773f1d05e21ef37342
3
+ size 59858847
train/fineweb2-sampled-decay-v2/lao_Laoo-tokenized-chunked-8192-512-32-backfill-nodups-sampled/000_00000-batch_0005-tokenized-chunked-8192-512-32-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0b451e6e8ae720551208f53bcbeeb33ef9c2364e58bdc341208f3cdfd31b30a
3
+ size 60557352
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0034-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcdba85c383bc884907710adbf998bfadd67058271b5f39023b228585cb8d8ec
3
+ size 16568411
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0058-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3446925451b08bc3660429e4696e8bc58efafc5e3f2138252cd68e34be74d992
3
+ size 14589894
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0062-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53d6d3e423457dbda55de6ae1af88d6e5cd8990796335c44a2deeab9a85879b1
3
+ size 15507493
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0074-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19b66155002edbf921dcdefeed18755f018490006137cd510a9f797eb33ca160
3
+ size 17180190
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0080-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd81aeda8903d8b75d367559320b95a69605b94f723898410ebb46c93cccfa20
3
+ size 15389923
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0100-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b17e04417d08345420afba912bc4bcab2ede551ae8efa41cefc65cf5f8fe459
3
+ size 15115746
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0162-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acca68f25df79a76376c6d6735d1e00b82fb4cbfe456a08e7a174a77163535e8
3
+ size 24655011
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0225-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fbce52f28045b8263b676fd7621ea6e27c6c14caa3df7b05c5cf44306d31c60
3
+ size 22771086
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0226-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c01d6c696e52c455cef4d0a0282cce1b865ed3efad6b4e9e4573adb334751e7
3
+ size 24260623
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0314-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106758, "hashes": {}}, "samples": 14694, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12334585, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18618242, "hashes": {}}, "samples": 5718, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 3356766, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0314-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de17a8652370601186ed8a589c04725b440717e9f79e96dfc451ce08b001009e
3
+ size 18618242
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0314-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0315-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bc4dc191315fb0a2da1618b18b9f7f962f039d688e30cf8b4974466ba77dcf3
3
+ size 14195752
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0322-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e31e06177e7805f3c0721f50da9749859164cb0484391be5d28a4801138931f
3
+ size 21845448
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0353-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68d8f59a8e401604efc3909a6b504fb41a218c51ccb8a0a2de6f4ad331faa427
3
+ size 15000232
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0457-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f05ac8aa6fbc6e2ec719511a569d7b88a49d2d791dd2ccedad3d8f110d693d5c
3
+ size 17232470
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0523-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:333ba0fa1d57696378c8d3ae5b064e10fd8ff3eb9098be069ac72f807dde6192
3
+ size 20962594
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0551-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41b431756c08f9989196ca033d43f236bd6575616dd619a9549175e5e998875b
3
+ size 28371179
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0566-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d91483b98d2eb3b19c3f892b8353d51e0a12be6968a44f588402262f1e3b0f2
3
+ size 24019435
train/fineweb2-sampled-decay-v2/slv_Latn_train-sampled/batch_0653-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c1cdc6d89668ced3d5a11b35591f7a98df4887cc18b6dd86a0c4833f9a5c8fb
3
+ size 37143994
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104350, "hashes": {}}, "samples": 13612, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 19907669, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 31862441, "hashes": {}}, "samples": 6671, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9569395, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 24417809, "total_tokens_skipped": 0, "percentiles": {"0th": 71, "10th": 203, "20th": 315, "30th": 444, "40th": 586, "50th": 766, "60th": 977, "70th": 1257, "80th": 1686, "90th": 2580, "95th": 3836, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0116-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0128-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103473, "hashes": {}}, "samples": 13697, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 21352038, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 32544014, "hashes": {}}, "samples": 6587, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10297095, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0128-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 24587992, "total_tokens_skipped": 0, "percentiles": {"0th": 72, "10th": 201, "20th": 315, "30th": 444, "40th": 591, "50th": 771, "60th": 987, "70th": 1277, "80th": 1731, "90th": 2620, "95th": 3875, "99th": 8190, "100th": 8191}}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0128-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108745, "hashes": {}}, "samples": 17083, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12930490, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12314988, "hashes": {}}, "samples": 3049, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 2342599, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96936eb7ea3c5c921d11bb0cea362353b4b6652e2b3d0e0b02c858b4e6a63928
3
+ size 12314988
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 19534559, "total_tokens_skipped": 0, "percentiles": {"0th": 68, "10th": 191, "20th": 283, "30th": 376, "40th": 477, "50th": 597, "60th": 772, "70th": 994, "80th": 1322, "90th": 2055, "95th": 3037, "99th": 6548, "100th": 8191}}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0206-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0264-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67102968, "hashes": {}}, "samples": 13386, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 13348100, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 36488081, "hashes": {}}, "samples": 6784, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7573223, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0264-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 25575526, "total_tokens_skipped": 0, "percentiles": {"0th": 67, "10th": 224, "20th": 384, "30th": 569, "40th": 768, "50th": 960, "60th": 1209, "70th": 1525, "80th": 1861, "90th": 2474, "95th": 3322, "99th": 6655, "100th": 8191}}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0264-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0356-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67103237, "hashes": {}}, "samples": 14705, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14232753, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 26841521, "hashes": {}}, "samples": 5456, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5711741, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0356-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 23164167, "total_tokens_skipped": 54, "percentiles": {"0th": 73, "10th": 217, "20th": 344, "30th": 480, "40th": 637, "50th": 808, "60th": 994, "70th": 1284, "80th": 1653, "90th": 2368, "95th": 3279, "99th": 7016, "100th": 8191}}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0356-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0372-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104670, "hashes": {}}, "samples": 13131, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14281750, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 40738934, "hashes": {}}, "samples": 7051, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8561379, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0372-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 26638391, "total_tokens_skipped": 0, "percentiles": {"0th": 74, "10th": 254, "20th": 406, "30th": 591, "40th": 790, "50th": 984, "60th": 1235, "70th": 1560, "80th": 1896, "90th": 2590, "95th": 3545, "99th": 7222, "100th": 8191}}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0372-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0403-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108224, "hashes": {}}, "samples": 14224, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14797961, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 32329626, "hashes": {}}, "samples": 5997, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7136550, "hashes": {}}}], "version": 2}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0403-tokenized-chunked-8192-512-32-backfill-nodups/stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"total_duplicated_tokens": 0, "total_tokens_written": 24536428, "total_tokens_skipped": 42, "percentiles": {"0th": 81, "10th": 228, "20th": 368, "30th": 520, "40th": 691, "50th": 867, "60th": 1074, "70th": 1367, "80th": 1753, "90th": 2443, "95th": 3345, "99th": 7983, "100th": 8191}}
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0403-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json ADDED
The diff for this file is too large to render. See raw diff
 
train/fineweb2-sampled-decay-v2/tha_Thai_train-sampled/batch_0415-tokenized-chunked-8192-512-32-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67079253, "hashes": {}}, "samples": 12360, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 14842295, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 32731986, "hashes": {}}, "samples": 7931, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6961749, "hashes": {}}}], "version": 2}