diff --git "a/lm_harness_eval.md" "b/lm_harness_eval.md" new file mode 100644--- /dev/null +++ "b/lm_harness_eval.md" @@ -0,0 +1,597 @@ +2025-06-11:04:05:16 INFO [__main__:440] Selected Tasks: ['arc_challenge', 'arc_easy', 'hellaswag', 'mmlu', 'openbookqa', 'piqa', 'pubmedqa', 'race', 'winogrande'] +2025-06-11:04:05:16 INFO [evaluator:189] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234 +2025-06-11:04:05:16 INFO [evaluator:227] Initializing hybrid model, with arguments: {'pretrained': '/home/mnt/mingyyan/checkpoints/Zebra-Llama-1B-8MLA-8Mamba-SFT'} +You are attempting to use Flash Attention 2.0 with a model not initialized on GPU. Make sure to move the model to GPU after initializing it on CPU with `model.to('cuda')`. +2025-06-11:04:05:16 WARNING [transformers.modeling_utils:329] You are attempting to use Flash Attention 2.0 with a model not initialized on GPU. Make sure to move the model to GPU after initializing it on CPU with `model.to('cuda')`. +LlamaForCausalLM( + (model): LlamaModel( + (embed_tokens): Embedding(128256, 2048) + (layers): ModuleList( + (0): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (1): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + (2): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (3): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + (4): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (5): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + (6): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (7): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + (8): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (9): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + (10): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (11): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + (12): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (13): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + (14): MLADecoderLayer( + (mla): DeepseekV3FlashAttention2( + (q_a_proj): Linear(in_features=2048, out_features=1344, bias=False) + (q_a_layernorm): Identity() + (q_b_proj): Linear(in_features=1344, out_features=2048, bias=False) + (kv_a_proj_with_mqa): Linear(in_features=2048, out_features=160, bias=False) + (kv_a_layernorm): Identity() + (kv_b_proj): Linear(in_features=128, out_features=768, bias=False) + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + (rotary_emb): DeepseekV3YarnRotaryEmbedding() + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): DeepseekV3RMSNorm() + (post_attention_layernorm): DeepseekV3RMSNorm() + ) + (15): Mamba2DecoderLayer( + (mamba): Mamba2( + (in_proj): Linear(in_features=2048, out_features=5152, bias=False) + (conv1d): Conv1d(3072, 3072, kernel_size=(4,), stride=(1,), padding=(3,), groups=3072) + (act): SiLU() + (norm): RMSNorm() + (out_proj): Linear(in_features=2048, out_features=2048, bias=False) + ) + (mlp): MLP( + (gate_proj): Linear(in_features=2048, out_features=8192, bias=False) + (up_proj): Linear(in_features=2048, out_features=8192, bias=False) + (down_proj): Linear(in_features=8192, out_features=2048, bias=False) + (act_fn): SiLU() + ) + (input_layernorm): RMSNorm() + (post_attention_layernorm): RMSNorm() + ) + ) + (norm): LlamaRMSNorm((2048,), eps=1e-05) + (rotary_emb): LlamaRotaryEmbedding() + ) + (lm_head): Linear(in_features=2048, out_features=128256, bias=False) +)2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of winogrande from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of race from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of pubmedqa from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of piqa from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of openbookqa from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_machine_learning from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_statistics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_physics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_mathematics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_computer_science from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_chemistry from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_biology from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_elementary_mathematics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_electrical_engineering from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_conceptual_physics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_computer_security from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_college_physics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_college_mathematics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_college_computer_science from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_college_chemistry from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_college_biology from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_astronomy from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_anatomy from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_abstract_algebra from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_virology from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_professional_medicine from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_professional_accounting from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_nutrition from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_miscellaneous from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_medical_genetics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_marketing from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_management from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_human_aging from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_global_facts from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_college_medicine from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_clinical_knowledge from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_business_ethics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_us_foreign_policy from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_sociology from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_security_studies from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_public_relations from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_professional_psychology from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_human_sexuality from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_psychology from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_microeconomics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_macroeconomics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_government_and_politics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_geography from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_econometrics from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_world_religions from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_professional_law from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_prehistory from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_philosophy from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_moral_scenarios from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_moral_disputes from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_logical_fallacies from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_jurisprudence from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_international_law from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_world_history from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_us_history from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_high_school_european_history from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of mmlu_formal_logic from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of hellaswag from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of arc_easy from None to 0 +2025-06-11:04:07:21 WARNING [evaluator:309] Overwriting default num_fewshot of arc_challenge from None to 0 +2025-06-11:04:07:21 INFO [api.task:434] Building contexts for winogrande on rank 0... + +#Params: 1267149568 + 0%| | 0/1267 [00:00