source ~/environments/clearml/bin/activate python /cache/git/research/automation/evaluation_scripts/queue_lm_evaluation_harness_vllm.py \ --model-id 0fe5857173ac484a89316214b14fcf96 \ --clearml-model \ --queue-name oneshot-a100x2 \ --project-name "LLM quantization - W8A16/llmcompressor/Qwen2.5" \ --task-name "Qwen2.5-72B-Instruct/openllm/vllm" \ --benchmark-tasks openllm \ --max-model-len 4096 \ --add-bos-token \ --batch-size auto \ --enable-chunked-prefill \ --gpu-memory-utilization 0.9 \ --max-num-batched-tokens 256 <