Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.  
							See raw diff
- train/math-sampled-decay/split_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_0-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
 - train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
 - train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_110-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_110-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_110-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
 - train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
 - train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
 - train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
 - train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
 - train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_167-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_167-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_167-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds +3 -0
 - train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds +3 -0
 - train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
 - train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_255-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_255-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_297-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/stats.json +1 -0
 - train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_334-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 - train/math-sampled-decay/split_334-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds +3 -0
 - train/math-sampled-decay/split_334-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json +0 -0
 - train/math-sampled-decay/split_338-tokenized-chunked-8192-512-32-backfill-nodups/index.json +1 -0
 
    	
        train/math-sampled-decay/split_0-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:047d7338d6644e8fcd5aa1289536b3f859e8d233df9be863fe85795a52c53eec
         
     | 
| 3 | 
         
            +
            size 60772565
         
     | 
    	
        train/math-sampled-decay/split_0-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106794, "hashes": {}}, "samples": 25756, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 19075075, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18970746, "hashes": {}}, "samples": 4244, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5786049, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:22b77fc02e1b217708645ca8934298cc555fba2d1396a1abc421a75f44ae1d8c
         
     | 
| 3 | 
         
            +
            size 67106794
         
     | 
    	
        train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:957bfa496a19a85a28687072c3d4bfe089131b9c431ad448d05af6511d32d417
         
     | 
| 3 | 
         
            +
            size 18970746
         
     | 
    	
        train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"total_duplicated_tokens": 0, "total_tokens_written": 21240664, "total_tokens_skipped": 0, "percentiles": {"0th": 34, "10th": 112, "20th": 114, "30th": 114, "40th": 434, "50th": 823, "60th": 928, "70th": 1029, "80th": 1150, "90th": 1349, "95th": 1543, "99th": 2021, "100th": 3068}}
         
     | 
    	
        train/math-sampled-decay/split_11-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_110-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 57549405, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12410384, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_110-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:5aec0ccbbdd84eb234cf52e999486ac03990270391348484fdba7e14e945fcfb
         
     | 
| 3 | 
         
            +
            size 57549405
         
     | 
    	
        train/math-sampled-decay/split_110-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 57861422, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12446449, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c8dd12a1238ad325721d06bab991ca11cde5fb5fec51c003f2b075a59ee8445d
         
     | 
| 3 | 
         
            +
            size 57861422
         
     | 
    	
        train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"total_duplicated_tokens": 0, "total_tokens_written": 14180763, "total_tokens_skipped": 0, "percentiles": {"0th": 142, "10th": 278, "20th": 326, "30th": 370, "40th": 409, "50th": 447, "60th": 488, "70th": 531, "80th": 589, "90th": 683, "95th": 800, "99th": 1079, "100th": 1210}}
         
     | 
    	
        train/math-sampled-decay/split_111-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 57877774, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12452532, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:30a56b5d6acaa64b2a82f81ad7988ab47eb141cc4668dde1c1f0e283214889bc
         
     | 
| 3 | 
         
            +
            size 57877774
         
     | 
    	
        train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"total_duplicated_tokens": 0, "total_tokens_written": 14184836, "total_tokens_skipped": 0, "percentiles": {"0th": 123, "10th": 279, "20th": 326, "30th": 371, "40th": 411, "50th": 448, "60th": 487, "70th": 531, "80th": 587, "90th": 682, "95th": 794, "99th": 1082, "100th": 1203}}
         
     | 
    	
        train/math-sampled-decay/split_112-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106799, "hashes": {}}, "samples": 20591, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 23075473, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 30542144, "hashes": {}}, "samples": 9409, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10501109, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:ef374578b960a011a92513feb7d76b5d41f32d0744df935a3104762d0fcebfed
         
     | 
| 3 | 
         
            +
            size 67106799
         
     | 
    	
        train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:442baf041668b1895ed8a1744490ca40ec56699fcca2a76945148a27d830e025
         
     | 
| 3 | 
         
            +
            size 30542144
         
     | 
    	
        train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"total_duplicated_tokens": 0, "total_tokens_written": 24052252, "total_tokens_skipped": 0, "percentiles": {"0th": 124, "10th": 497, "20th": 575, "30th": 635, "40th": 690, "50th": 748, "60th": 811, "70th": 892, "80th": 999, "90th": 1190, "95th": 1384, "99th": 1793, "100th": 2474}}
         
     | 
    	
        train/math-sampled-decay/split_120-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 58033568, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 12485075, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:06041b9a11021792b7febf4f7ddffb02c505bb04e7ce51024ecdd2784b2399df
         
     | 
| 3 | 
         
            +
            size 58033568
         
     | 
    	
        train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"total_duplicated_tokens": 0, "total_tokens_written": 14223794, "total_tokens_skipped": 0, "percentiles": {"0th": 130, "10th": 280, "20th": 327, "30th": 370, "40th": 410, "50th": 449, "60th": 490, "70th": 534, "80th": 589, "90th": 683, "95th": 799, "99th": 1083, "100th": 1219}}
         
     | 
    	
        train/math-sampled-decay/split_126-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_167-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 54927656, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11767599, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_167-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:cad1cf108198c600a0d1a462f14144b9e0c3b729d4eb0d9ecdc4a0d3a520b651
         
     | 
| 3 | 
         
            +
            size 54927656
         
     | 
    	
        train/math-sampled-decay/split_167-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67104827, "hashes": {}}, "samples": 15105, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 20460667, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 67099362, "hashes": {}}, "samples": 12573, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 20636508, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00002.mds", "bytes": 56872471, "hashes": {}}, "samples": 3373, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00002.mds.zstd", "bytes": 17947150, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:b8e89dbc6aba114cd1e4712175e845c934e161911edc01aefd69fef51c32b069
         
     | 
| 3 | 
         
            +
            size 67104827
         
     | 
    	
        train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/shard.00001.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:5282bd21176accda1caef9951ecf57fe342c6159d7cc83219214f0890f0763d5
         
     | 
| 3 | 
         
            +
            size 67099362
         
     | 
    	
        train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/shard.00002.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:84a65e2efe5181d45089af1f41be618b446c42c38424a396d943f7221bed5f93
         
     | 
| 3 | 
         
            +
            size 56872471
         
     | 
    	
        train/math-sampled-decay/split_238-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 54300572, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11603740, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:b031e91d3f1e26c362ddb8112319dbd5b4c09c17f05b2abb79576dbf196f79ec
         
     | 
| 3 | 
         
            +
            size 54300572
         
     | 
    	
        train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"total_duplicated_tokens": 0, "total_tokens_written": 13290833, "total_tokens_skipped": 0, "percentiles": {"0th": 112, "10th": 256, "20th": 301, "30th": 341, "40th": 381, "50th": 418, "60th": 457, "70th": 499, "80th": 551, "90th": 645, "95th": 749, "99th": 1078, "100th": 1230}}
         
     | 
    	
        train/math-sampled-decay/split_248-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_255-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 54269697, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11604134, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_255-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:94356f339848d21f374c12880c155c4c2a6e05b40c16cce25148c9883f6006fe
         
     | 
| 3 | 
         
            +
            size 54269697
         
     | 
    	
        train/math-sampled-decay/split_297-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:5b3745c07e548784d589f341e73dd3981f0ee99ad6baa37b58351cd55ea90526
         
     | 
| 3 | 
         
            +
            size 57716704
         
     | 
    	
        train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 54365996, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11640266, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:ab8f812a0a62cb08310e793240336f8a9f4deac3b712be34f9bcfc82603a1610
         
     | 
| 3 | 
         
            +
            size 54365996
         
     | 
    	
        train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/stats.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"total_duplicated_tokens": 0, "total_tokens_written": 13307169, "total_tokens_skipped": 0, "percentiles": {"0th": 122, "10th": 257, "20th": 301, "30th": 342, "40th": 382, "50th": 419, "60th": 459, "70th": 501, "80th": 553, "90th": 646, "95th": 744, "99th": 1074, "100th": 1185}}
         
     | 
    	
        train/math-sampled-decay/split_312-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_334-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 54110033, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11561917, "hashes": {}}}], "version": 2}
         
     | 
    	
        train/math-sampled-decay/split_334-tokenized-chunked-8192-512-32-backfill-nodups/shard.00000.mds
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:b3a0f644395a14bbfd06e3398574905615ef48577c10c6594949cf0b18e94588
         
     | 
| 3 | 
         
            +
            size 54110033
         
     | 
    	
        train/math-sampled-decay/split_334-tokenized-chunked-8192-512-32-backfill-nodups/token_decile.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        train/math-sampled-decay/split_338-tokenized-chunked-8192-512-32-backfill-nodups/index.json
    ADDED
    
    | 
         @@ -0,0 +1 @@ 
     | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {"shards": [{"column_encodings": ["str", "ndarray:uint32"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 54454439, "hashes": {}}, "samples": 30000, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 11629708, "hashes": {}}}], "version": 2}
         
     |