source ~/environments/clearml/bin/activate python /cache/git/research/automation/evaluation_scripts/queue_lm_evaluation_harness_vllm.py \ --model-id 0909a970c53644f593c62b2e076ee763 \ --clearml-model \ --queue-name oneshot-a100x1 \ --project-name "LLM quantization - W8A8/llmcompressor/Qwen2.5" \ --task-name "Qwen2.5-7B/openllm/vllm" \ --benchmark-tasks openllm \ --max-model-len 4096 \ --add-bos-token \ --batch-size auto \ --enable-chunked-prefill \ --gpu-memory-utilization 0.9 \ --max-num-batched-tokens 256 python /cache/git/research/automation/evaluation_scripts/queue_lm_evaluation_harness_vllm.py \ --model-id 0323219f7a8b4e938c2583ab17851b91 \ --clearml-model \ --queue-name oneshot-a100x2 \ --project-name "LLM quantization - W8A8/llmcompressor/Qwen2.5" \ --task-name "Qwen2.5-72B/openllm/vllm" \ --benchmark-tasks openllm \ --max-model-len 4096 \ --add-bos-token \ --batch-size auto \ --enable-chunked-prefill \ --gpu-memory-utilization 0.9 \ --max-num-batched-tokens 256 python /cache/git/research/automation/evaluation_scripts/queue_lm_evaluation_harness_vllm.py \ --model-id 31c634f743fe458687556bc85a6aeb32 \ --clearml-model \ --queue-name oneshot-a100x2 \ --project-name "LLM quantization - W8A8/llmcompressor/Qwen2.5" \ --task-name "Qwen2.5-72B-Instruct/openllm/vllm" \ --benchmark-tasks openllm \ --max-model-len 4096 \ --add-bos-token \ --batch-size auto \ --enable-chunked-prefill \ --gpu-memory-utilization 0.9 \ --max-num-batched-tokens 256 <